Question about the relationship between a project and Mobile Me album

Hello esteemed members of the Aperture community
I have a puzzling issue. I have published a project on MobileMe. My project is arranged somewhat chronologically, and I want MobileMe to follow this. However, the MobileMe album in Aperture does not synchronize with the project, and subsequently, the gallery on MobileMe does neither. How do I tell the MobileMe album in Aperture to synchronize itself with my project, and update changes as I edit the project, or am I just totally confused...always a possibility?
Thanks in advance

You sort of answered your own question in the title: Your Mobile Me gallery is an album. By design, changing the sort order of a project does not affect the sort order of related albums and vice versa. You'll have to make those changes in the album to have them reflected in the gallery.
I know of no way to have an album synchronize with a project. (Note: if you make adjustments to images or delete images from a project, those changes will be reflected in the album, but the sort order will not.)
DLS

Similar Messages

  • Question the relationship between AR invoice and Outgoing payment

    Dear All,
    Does anyone the know the the relationship between AR invoice and Outgoing payment? because I would like to print the report in specific date range. For example, today is 02 July 2009. I want to print "AR invoice as at 31 Mar 2009", by using SQL. It seems like date back.
    From Samson

    Hi Samson,
                      Even i haven't done much r n d  bout this but what i came to know is that, there might be a scenario where you may want to make a payment to your customer. In this case we can use the Outgoing payment for the same. The system doesnt stop you from doin so. I have seen this in SAP B1 2005B, PL34.
    Regards,
    Joseph

  • How to create the relationship between ESSBASE 11 and DM in OBIEE11

    Hi Experts,
    I have one requirement that there is one property table named 'Store Master' in DW,and it contains a lot of attribute, such as Open Date, Close Date, IS 24 Hour etc.
    But another data source is essbase and based on this source, I create all reports.
    In ESSBASE, it has one dimension and hierarchy Location, and it has four level, Country(L1),Region (L2),Province(L3),Store(L4)
    So I want to know how to create the relationship between Location (ESSBASE) and Store Master (DM).
    I try to create one relationship in physical layer between Gen4,Location and Store, then drag the open date and close date into Location Dimension in BMM,then Presentation Layer.
    When I drag column 'Open Date' ,'Gen4,Location ' and 'Sales' into reports, it will generate the following error message:
    State: HY000. Code: 10058. [NQODBC] [SQL_STATE: HY000] [nQSError: 10058] A general error has occurred. [nQSError: 43113] Message returned from OBIS. [nQSError: 43119] Query Failed: [nQSError: 14020] None of the fact tables are compatible with the query request Dim Region.Store Open Date. (HY000)
    However, when I remove the column 'Open Date'. it will be ok.
    So what I missing the steps? Please help me. Thanks.

    >
    '2. Now, pull the 'Store' column from relational DB onto the Gen5, Location column from Essbase. This action now creates, two logical sources for your 'Store' column.'
    If the length from different data source is not same,such as 1001(DM),L_1001(ESSBASE), can I drag the 'Store' column from relational DB onto the Gen5, Location column from Essbase?
    I think it does not work.Right?Hi,
    I am not sure if you are talking about the length(as in varchar(128)) of the member value being different in different sources, or the member itself is different in both the sources.
    I am still assuming, that you are referring to the members not same in both the sources.If it is, the whole concept of federation is based on conforming dimensions. So, it needs that the same dimension information is present in both the sources and then only, you know we can analyze the numbers based on this dimension. So, either the dimension being different in both sources, or the members not present in both the dimensions might lead to incorrect numbers.
    So I select Store Attributes in relation DB and Location in ESSBASE in physical layer, then create the physical join, such as right("Hour Sales"."H_Sales".""."H_Sales"."Gen6,Location",4) = "Authorization".""."EDW"."T_EDW_MDM_STORE"."US_CODE", then drag the OPEN_DATE and CLOSE_DATE in relation DB to Location in ESSBASE in BMM,finially drag them into presentation layer.We create physical layer relationships, to send over the same relation to the underlying database during querying. So, creating a physical relationship between essbase cube and relation database would not help here.
    When you set up this federation, BI Server sends individual queries to each source and maps the conforming dimension members internally.
    Hope I was clear, and this helps.
    Thank you,
    Dhar

  • How to create the relationship between ESSBASE 11 and DM  in OBIEE 11G?

    Hi Experts,
    I have one requirement that there is one property table named 'Store Master' in DW,and it contains a lot of attribute, such as Open Date, Close Date, IS 24 Hour etc.
    But another data source is essbase and based on this source, I create all reports.
    In ESSBASE, it has one dimension and hierarchy Location, and it has four level, Country(L1),Region (L2),Province(L3),Store(L4)
    So I want to know how to create the relationship between Location (ESSBASE) and Store Master (DM).
    I try to create one relationship in physical layer between Gen4,Location and Store, then drag the open date and close date into Location Dimension in BMM,then Presentation Layer.
    When I drag column 'Open Date' ,'Gen4,Location ' and 'Sales' into reports, it will generate the following error message:
    State: HY000. Code: 10058. [NQODBC] [SQL_STATE: HY000] [nQSError: 10058] A general error has occurred. [nQSError: 43113] Message returned from OBIS. [nQSError: 43119] Query Failed: [nQSError: 14020] None of the fact tables are compatible with the query request Dim Region.Store Open Date. (HY000)
    However, when I remove the column 'Open Date'. it will be ok
    So what I missing the steps? Please help me. Thanks.

    >
    '2. Now, pull the 'Store' column from relational DB onto the Gen5, Location column from Essbase. This action now creates, two logical sources for your 'Store' column.'
    If the length from different data source is not same,such as 1001(DM),L_1001(ESSBASE), can I drag the 'Store' column from relational DB onto the Gen5, Location column from Essbase?
    I think it does not work.Right?Hi,
    I am not sure if you are talking about the length(as in varchar(128)) of the member value being different in different sources, or the member itself is different in both the sources.
    I am still assuming, that you are referring to the members not same in both the sources.If it is, the whole concept of federation is based on conforming dimensions. So, it needs that the same dimension information is present in both the sources and then only, you know we can analyze the numbers based on this dimension. So, either the dimension being different in both sources, or the members not present in both the dimensions might lead to incorrect numbers.
    So I select Store Attributes in relation DB and Location in ESSBASE in physical layer, then create the physical join, such as right("Hour Sales"."H_Sales".""."H_Sales"."Gen6,Location",4) = "Authorization".""."EDW"."T_EDW_MDM_STORE"."US_CODE", then drag the OPEN_DATE and CLOSE_DATE in relation DB to Location in ESSBASE in BMM,finially drag them into presentation layer.We create physical layer relationships, to send over the same relation to the underlying database during querying. So, creating a physical relationship between essbase cube and relation database would not help here.
    When you set up this federation, BI Server sends individual queries to each source and maps the conforming dimension members internally.
    Hope I was clear, and this helps.
    Thank you,
    Dhar

  • In which table we can find the relationship between Role id and Task id

    Hi Experts,
    In which table we can find the relationship between Role id and Task id in Cprojects.
    Thanks
    Subhaskar

    Hi Subhaskar,
    Apart from DPR_ENTITY_LINK , you can also get it from table DPR_PART.
    Please go through the below link
    http://wiki.sdn.sap.com/wiki/display/PLM/cProjectstablesin+SAP
    I hope this will help you.
    Regards,
    Rahul

  • The relationship between production order and purchase requistion After running MPS

    Dear Experts:
         I met a issue . My users want to know the relationship between production order and purchase requistion . In my company we create production order and  purchase req. by  running MRP. our operations are:
    1. T-CODE: MD41  create planned order .
    2. T-CODE: CO41 convert planned order to production order .
    3. T-CODE: MD15 convert planned order to purchase Requisition.
      but I can't  find the  relation of them. would you like to give me some advices? thank you very much.

    Hello
    On the standard MTS process there is no link between the requirement (STO) and the replenishment element (production order).
    See the below extraction from note 12955:
    There is no index that the system could use to determine which issue element (sales order, planning, reservation, dependent requirement, ...) is covered by a certain receipt element (purchase order, production order, purchase requisition, planned order, stock level ...).
    A production order, for example, can be created to cover several different requirements, even when you are using lot sizing procedure EX.
    You can use transaction MD09 or MD04to determine the pegged requirements, but they are calculated dynamically and there is not fixed link between requirement and receipt.
    BR
    Caetano

  • What is the relationship between CGI-Executables and cgi-bin?

    I am trying to get CGIs working locally on my Mac and am having difficulty understanding the relationship between CGI-Executables/ and cgi-bin/.
    I have activated CGI in my httpd.conf file, and have the following perl test script: /Library/WebServer/CGI-Executables/test.cgi file, and have the file permissions set appropriatly
    If I call the script from the actual url: http://127.0.0.1/CGI-Executables/test.cgi, it doesn't work. If I use: http://127.0.0.1/cgi-bin/test.cgi, it works fine.
    Why cgi-bin and not CGI-Executables? What is the relationship between /Library/WebServer/CGI-Executables/ and cgi-bin?
    And how do I get CGIs working in individual users' Sites/ file? Do all CGIs go in the /Library/WebServer/CGI-Executables/ directory, or can each users' Site/ directory have it's own cgi-bin?
    Ti Powerbook G4   Mac OS X (10.4.4)  

    Now I am having trouble getting CGI to work in the user directories. I have a perl script that works for 127.0.0.1/cgi-bin/test.cgi, but not for 127.0.0.1/~joe/cgi-bin/test.cgi.
    My apache error log says:
    [Tue Feb 14 08:14:23 2006] [error] [client 127.0.0.1] Options ExecCGI is off in this directory: /Users/joe/Sites/cgi-bin/first.cgi
    Permissions for Library/WebServer/CGI-Exectuables/test.cgi and /Users/joe/Sites/cgi-bin/ are 777.
    I think the problem is somewhere in my directory configs. Apache has so many, it's hard to know what's what.
    Here's the first directory directives in httpd.conf:
    # First, we configure the "default" to be a very restrictive set of
    # permissions.
    <Directory />
    Options FollowSymLinks
    AllowOverride Options
    </Directory>
    Farther down, there's this, which I added ExecCGI and index.cgi:
    # Control access to UserDir directories. The following is an example
    # for a site where these directories are restricted to read-only.
    <Directory /Users/*/Sites>
    AllowOverride FileInfo AuthConfig Limit
    Options MultiViews Indexes SymLinksIfOwnerMatch Includes ExecCGI
    DirectoryIndex index.html index.cgi
    <Limit GET POST OPTIONS PROPFIND>
    Order allow,deny
    Allow from all
    </Limit>
    <LimitExcept GET POST OPTIONS PROPFIND>
    Order deny,allow
    Deny from all
    </LimitExcept>
    </Directory>
    Then down at the bottom of the file is something I suspect was added automatically when the user account was created:
    Include /private/etc/httpd/users/*.conf
    <Directory "/Users/joe/Sites/">
    Options Indexes MultiViews Includes
    AllowOverride None
    Order allow,deny
    Allow from all
    </Directory>
    To further confuse things, there is another file /etc/httpd/user/joe.conf that has additional user-directory directives:
    <Directory "/Users/joe/Sites/">
    Options Indexes MultiViews
    AllowOverride None
    Order allow,deny
    Allow from all
    </Directory>
    Is there something in all these directory configurations that is preventing the user from accessing cgi-bin? Can I add directives to an .htaccess file in the user directory.
    Ti Powerbook G4   Mac OS X (10.4.4)  

  • What is the Relationship between Update Rule and Infopackage/Transfer Rules

    Hello Experts,
    can you help explain these scenarios to me:
    What is the relationship between update rules and infopackges?
    What is the relationship between transfer rules and infopackges?
    What is the relationship between update rules and transfer rules?

    Hello,
    Info-packages trigger the data load into BW system from the source system. An info-package is created for a specific data source and info source combination.
    1) When you trigger the info-package it sends an request to the source system to start the extraction process and this corresponds to the step "Data Request Received" in monitor under details tab. It is the first step.
    2) Then the selections specified in the info-package are transferred to the source system to select the requested data from BW system. Then it is sent back to BW system.
    3) Once it reach BW, it gets into the PSA and then it gets updated into the data target (depends on what settings are made in IP). No rules (Transfer rules / update rules) comes into picture when data is stored in PSA. Only the transfer structure is used to map the BW info-objects with the source system fields.
    4) After data is stored into PSA, the update into the data targets begins. First the transfer rules are applied -> data then flows through the communication structure -> update rules are applied and finally the records gets updated into the data target. 
    The extraction flow is: -
    Source System (DS) -> PSA -> Transfer Rules -> Communication Structure -> Update Rules -> Data target.
    Info-package actually triggers the whole data flow in BW system.
    Hope it gives you a clear picture.
    Thx,
    Soumya

  • About the Differences between MyServices Professional and MyServices Consumer

    About the Differences between MyServices Professional and MyServices Consumer
    MyServices Professional is different from MyServices Consumer in the following ways:
    There are Common Tasks links on the home page for Order on Behalf and for Authorizations (if the Authorization portlet is enabled in the user's profile).
    Automatically grants the Order on Behalf permission for all people in the Home OU of the MyServices Professional user.
    Adds the "Authorizations for Others" view on the Authorizations page. This view allows the MyServices Professional user to view -- but not perform -- all Authorizations for people in their Home OU.
    See the attached file for more detail and screenshots of: MyServices Consumer with and without the Order on Behalf link displayed, and Requisition and Authorization list views; MyServices Professional Home page with and without the Authorizations portlet visible, and Requisition and Authorization list views. 
    Note: RequestCenter 2006 cannot be configured to automatically add users to MyServices Professional on creation of their user record via directory integration. This can be accomplished in RC2007, since roles can be imported from LDAP in that version.

    Hi Koji,
    One thing to consider is the number of different source types. BO supports SAP and all the relational DBs. There is also the fact that BO provides a single front end to different reporting type which makes the user experience much better. Finally, users ability to build there own reports, without a high level of traiining is one of the big selling points for BO.
    Here is a good overview  of the combination of the 2 , including some roadmap
    https://www.sdn.sap.com/irj/scn/go/portal/prtroot/docs/library/uuid/10c3bca6-7dbc-2a10-7aa8-81d2731c7bb1
    Hope this helps
    Alan

  • Some questions about the integration between BIEE and EBS

    Hi, dear,
    I'm a new bie of BIEE. In these days, have a look about BIEE architecture and the BIEE components. In the next project, there are some work about BIEE development based on EBS application. I have some questions about the integration :
    1) generally, is the BIEE database and application server decentralized with EBS database and application? Both BIEE 10g and 11g version can be integrated with EBS R12?
    2) In BIEE administrator tool, the first step is to create physical tables. if the source appliation is EBS, is it still needed to create the physical tables?
    3) if the physical tables creation is needed, how to complete the data transfer from the EBS source tables to BIEE physical tables? which ETL tool is prefer for most developers? warehouse builder or Oracle Data Integration?
    4) During data transfer phase, if there are many many large volume data needs to transfer, how to keep the completeness? for example, it needs to transfer 1 million rows from source database to BIEE physical tables, when 50%is completed, the users try to open the BIEE report, can they see the new 50% data on the reports? is there some transaction control in ETL phase?
    could anyone give some guide for me? I'm very appreciated if you can also give any other information.
    Thanks in advance.

    1) generally, is the BIEE database and application server decentralized with EBS database and application? Both BIEE 10g and 11g version can be integrated with EBS R12?You, shud consider OBI Application here which uses OBIEE as a reporting tool with different pre-built modules. Both 10g & 11g comes with different versions of BI apps which supports sources like Siebel CRM, EBS, Peoplesoft, JD Edwards etc..
    2) In BIEE administrator tool, the first step is to create physical tables. if the source appliation is EBS, is it still needed to create the physical tables?Its independent of any soure. This is OBIEE modeling to create RPD with all the layers. If you build it from scratch then you will require to create all the layers else if BI Apps is used then you will get pre-built RPD along with other pre-built components.
    3) if the physical tables creation is needed, how to complete the data transfer from the EBS source tables to BIEE physical tables? which ETL tool is prefer for most developers? warehouse builder or Oracle Data Integration?BI apps comes with pre-built ETL mapping to use with the tools majorly with Informatica. Only BI Apps 7.9.5.2 comes with ODI but oracle has plans to have only ODI for any further releases.
    4) During data transfer phase, if there are many many large volume data needs to transfer, how to keep the completeness? for example, it needs to transfer 1 million rows from source database to BIEE physical tables, when 50%is completed, the users try to open the BIEE report, can they see the new 50% data on the reports? is there some transaction control in ETL phase?User will still see old data because its good to turn on Cache and purge it after every load.
    Refer..http://www.oracle.com/us/solutions/ent-performance-bi/bi-applications-066544.html
    and many more docs on google
    Hope this helps

  • The Question about stock transfer between HU-Management and WM-Management

    Hi,
    There is a scenario about stock transfer between HU-Management and WM-Management storage location. I use transaction MB1B , movement type 313 , 315. After Good issue from WM-management storage location, outbound delivery will genarated, then Pack, Create/Confirm TO, at last post goods issue for the outboud delivery. But when i do movement type 315, there is a warning message "Data of preceding document was not transmitted", and from the F1 help i find this system reponse "You can maintain an indicator that makes information about preceding documents in this delivery available under delivery type in Customizing. For some characteristics of this indicator, the type of preceding document and the related document and item numbers must be transmitted to delivery creation. At least one of these parameters is missing.".
    So, My questions are:
    1. Generallily, Outbound delivery is created by SO, inbound delivery is created referenc PO, but how the stock transfer for 2 steps generate the outbould delivery and inbound delivery? Could you pls tell me the where i can config this in the IMG?
    2. What's "Data of preceding document was not transmitted" mean ? how to fix this issue ?
    Best Regards
    Boxer Du
    I am the SAP fans, focus on MM and WM. I am interesting TRM Yard Management and Cross Docking now.Very Gladly to talk you about these areas. I want to exchange the knowledge with you, and want to be a good friend of you. Pls contact me. You can find My MSN in the profile. Thanks.

    Hi,
    Sure, The inbound delivery type is set up in the IMG->Logistics General -> HU management -> Basics ->Delivery type -> Delivery type determination.
    For Inbound delivery type 'HID' is maintained in this view.
    Best Regards.
    Some One want to discuss the details , can contact me. Thanks.

  • Questions about the future directions of InfoPath and Microsoft Access

    Because I have been confused about the future directions of InfoPath and Access, I wrote a short blog post, which I have pasted below.  My question is, can Microsoft provide some more guidance on Access vs InfoPath?  If Access is indeed the
    the future direction, can you please provide some pointers to resources that can help us transition.
    A Short Review of MS Access,  SharePoint ,InfoPath (2013)
    After the November 2012 SharePoint Conference in Las Vegas (#spc12) there were a number of blog posts and people saying that Access 2013 was great and would be replacing InfoPath.
    Marc Anderson and
    Ruven
    Gotz have short posts summarizing #SPC12 and touch on the subject of InfoPath and Access.  I also spoke with several individuals shortly after #SPC12 and they were excited about Access 2013.
    Since I rely heavily on InfoPath for much of what I do, I wanted to kick the tires and see what everyone was talking about.  First, a few notes about SharePoint 2013 and Office 2013.
    Microsoft should rename Access:  If you have tried to
    bing the word Access, then you know that the search results are overwhelming,  because the word "access" is used for so many other purposes.
    Use the RTM version of Office 2013! I was performing  my investigations by leveraging the Office 365 Beta and the Office Applications that come with it.  Not much was working!  After I uninstalled the Beta
    version of Office 2013 and Installed the RTM version everything seemed to work. 
    Windows 7 and IE 9.0: The drag & drop that is being touted as
    "manna from heaven" seems to work well with Firefox, Safari, Chrome, and IE 10, but it does
    not work with IE 9!
    Non Microsoft browsers are still second class citizens (except for the drag & drop):  Microsoft has finally done away with the Active X's for datasheet view (that is brilliant!)  One can now see the newly
    minted datasheet view in many browsers.  But I still find that one needs to be in IE to do a number of key manual opeartions.  For example pasting numerous rows of data from a spreadsheet into the new sheet view works only in IE.
    Access 2013
    Microsoft Access is touted as a non developer tool.  I know a number of people who are on the business side and love Access.  These people will like the renewed energy that Microsoft has applied to Access.   One should note that conceptually
    this functionality existed  in SharePoint 2010.  Here is what I see that is new:
    +The new version works better:Although I haven't tested it fully, it seems that many more features that work on the Desktop version of Access, now work in the published Web Application which is hosted by SharePoint 2013.
    +The new version has fairly robust browser based forms: End users can now fill in data using forms that are rendered in the browser.
    +The published version that resides in SharePoint now uses its own SQL server tables.  So the engine appears to be SQL Server, not a modified version of Access built for a server.  This stands to be quite robust, but I imagine
    that some quirks will surface due to new model.
    The limitations are:
    - Access is pretty much its own application and does not integrate with SharePoint building blocks.  For example, an Access table is not easy to integrate with a SharePoint workflow.  Further, although  Access forms now are visible
    in a browser, they are not nearly as rich as InfoPath forms.  InfoPath forms cannot easily integrate with the Access tables.  In order to integrate SharePoint features and Access, one needs to tie SharePoint list(s) to Access as data sources. 
    This overcomplicates the model and one may very well run into synchronization issues.
    In short, although I believe there will be some people who will be happy with newly minted Access, I do not think that Access will be able to replace InfoPath and Workflow technologies.  I do agree that Microsoft does not seem to have put much energy into
    InfoPath.  If they leave a vacuum in this space, then others may very well take over, for instance Nintex Forms.

    Although I agree with the comments made by Ruven and Marc after the conference please note that those blog posts are their own personal opinions and that they don't work for Microsoft.  A lot of people have asked Microsoft to clarify the future of InfoPath
    and the response up until now has been a deafening silence.  In my opinion speculations about Access replacing InfoPath are premature.
    I should also point out that most of your comments about Access 2013 above are actually about Access Services in SharePoint Server 2013.  Access Services are NOT Access they are a shared service offering in SharePoint that convert and publish Access
    databases and forms.  As you point out once an Access database is published through Access services it is converted to SQL Server and no longer resides in Access.  Of course I don't work for Microsoft either so these are all just my personal opinion. 
    Having said that I doubt you will get the confirmation/response that you are hoping for from any official Microsoft channel.
    Paul Stork SharePoint Server
    MVP Principal Solutions Architect: BlueChip Consulting Group
    Blog: http://dontpapanic.com/blog
    Twitter: Follow @pstork
    Please remember to mark your question as "answered" if this solves your problem.

  • Question about main difference between Java bean and Java class in JSP

    Hi All,
    I am new to Java Bean and wonder what is the main difference to use a Bean or an Object in the jsp. I have search on the forum and find some post also asking the question but still answer my doubt. Indeed, what is the real advantage of using bean in jsp.
    Let me give an example to illustrate my question:
    <code>
    <%@ page errorPage="errorpage.jsp" %>
    <%@ page import="ShoppingCart" %>
    <!-- Instantiate the Counter bean with an id of "counter" -->
    <jsp:useBean id="cart" scope="session" class="ShoppingCart" />
    <html>
    <head><title>Shopping Cart</title></head>
    <body bgcolor="#FFFFFF">
    Your cart's ID is: <%=cart.getId()%>.
    </body>
    <html>
    </code>
    In the above code, I can also create a object of ShoppingCart by new operator then get the id at the following way.
    <code>
    <%
    ShoppingCart cart = new ShoppingCart();
    out.println(cart.getId());
    %>
    </code>
    Now my question is what is the difference between the two method? As in my mind, a normal class can also have it setter and getter methods for its properties. But someone may say that, there is a scope="session", which can be declared in an normal object. It may be a point but it can be easily solved but putting the object in session by "session.setAttribute("cart", cart)".
    I have been searching on this issue on the internet for a long time and most of them just say someting like "persistance of state", "bean follow some conventions of naming", "bean must implement ser" and so on. All of above can be solved by other means, for example, a normal class can also follow the convention. I am really get confused with it, and really want to know what is the main point(s) of using the java bean.
    Any help will be highly apprecaited. Thanks!!!
    Best Regards,
    Alex

    Hi All,
    I am new to Java Bean and wonder what is the main
    difference to use a Bean or an Object in the jsp. The first thing to realize is that JavaBeans are just Plain Old Java Objects (POJOs) that follow a specific set of semantics (get/set methods, etc...). So what is the difference between a Bean and an Object? Nothing.
    <jsp:useBean id="cart" scope="session" class="ShoppingCart" />
    In the above code, I can also create a object of
    ShoppingCart by new operator then get the id at the
    following way.
    ShoppingCart cart = new ShoppingCart();
    out.println(cart.getId());
    ...Sure you could. And if the Cart was in a package (it has to be) you also need to put an import statement in. Oh, and to make sure the object is accessable in the same scope, you have to put it into the PageContext scope. And to totally equal, you first check to see if that object already exists in scope. So to get the equivalant of this:
    <jsp:useBean id="cart" class="my.pack.ShoppingCart"/>Then your scriptlet looks like this:
    <%@ page import="my.pack.ShoppingCart %>
    <%
      ShoppingCart cart = pageContext.getAttribute("cart");
      if (cart == null) {
        cart = new ShoppingCart();
        pageContext.setAttribute("cart", cart);
    %>So it is a lot more work.
    As in my mind, a normal class can also
    have it setter and getter methods for its properties.True ... See below.
    But someone may say that, there is a scope="session",
    which can be declared in an normal object.As long as the object is serializeable, yes.
    It may be
    a point but it can be easily solved but putting the
    object in session by "session.setAttribute("cart",
    cart)".Possible, but if the object isn't serializable it can be unsafe. As the point I mentioned above, the useBean tag allows you to check if the bean exists already, and use that, or make a new one if it does not yet exist in one line. A lot easier than the code you need to use otherwise.
    I have been searching on this issue on the internet
    for a long time and most of them just say someting
    like "persistance of state", "bean follow some
    conventions of naming", "bean must implement ser" and
    so on. Right, that would go along the lines of the definition of what a JavaBean is.
    All of above can be solved by other means, for
    example, a normal class can also follow the
    convention. And if it does - then it is a JavaBean! A JavaBean is any Object whose class definition would include all of the following:
    1) A public, no-argument constructor
    2) Implements Serializeable
    3) Properties are revealed through public mutator methods (void return type, start with 'set' have a single Object parameter list) and public accessor methods (Object return type, void parameter list, begin with 'get').
    4) Contain any necessary event handling methods. Depending on the purpose of the bean, you may include event handlers for when the properties change.
    I am really get confused with it, and
    really want to know what is the main point(s) of
    using the java bean.JavaBeans are normal objects that follow these conventions. Because they do, then you can access them through simplified means. For example, One way of having an object in session that contains data I want to print our might be:
    <%@ page import="my.pack.ShoppingCart %>
    <%
      ShoppingCart cart = session.getAttribute("cart");
      if (cart == null) {
        cart = new ShoppingCart();
        session.setAttribute("cart", cart);
    %>Then later where I want to print a total:
    <% out.print(cart.getTotal() %>Or, if the cart is a JavaBean I could do this:
    <jsp:useBean id="cart" class="my.pack.ShoppingCart" scope="session"/>
    Then later on:
    <jsp:getProperty name="cart" property="total"/>
    Or perhaps I want to set some properties on the object that I get off of the URL's parameter group. I could do this:
    <%
      ShoppingCart cart = session.getAttribute("cart");
      if (cart == null) {
        cart = new ShoppingCart();
        cart.setCreditCard(request.getParameter("creditCard"));
        cart.setFirstName(request.getParameter("firstName"));
        cart.setLastName(request.getParameter("lastName"));
        cart.setBillingAddress1(request.getParameter("billingAddress1"));
        cart.setBillingAddress2(request.getParameter("billingAddress2"));
        cart.setZipCode(request.getParameter("zipCode"));
        cart.setRegion(request.getParameter("region"));
        cart.setCountry(request.getParameter("country"));
        pageContext.setAttribute("cart", cart);
        session.setAttribute("cart", cart);
      }Or you could use:
    <jsp:useBean id="cart" class="my.pack.ShoppingCart" scope="session">
      <jsp:setProperty name="cart" property="*"/>
    </jsp:useBean>The second seems easier to me.
    It also allows you to use your objects in more varied cases - for example, JSTL (the standard tag libraries) and EL (expression language) only work with JavaBeans (objects that follow the JavaBeans conventions) because they expect objects to have the no-arg constuctor, and properties accessed/changed via getXXX and setXXX methods.
    >
    Any help will be highly apprecaited. Thanks!!!
    Best Regards,
    Alex

  • Question about the difference between HDDs and SSDs

    Hi. I currently have a MacBook (Black one) that's overheating a lot for the past few months. I'm thinking about purchasing a MacBook Pro for school because it seems to be more stable. I have a question though. What is the difference between HDDs and SSDs? Which one is better? All I know is that with my MacBook and my iBook is that I had hard drive failures (iBook hard drive clicking, MacBook hard drive sounding loud and making my computer overheat). Are hard drive failures normal with Macs?
    Thank you

    astoldbywesley wrote:
    Is a 128GB Solid State Drive better than a 500GB Serial ATA Drive @ 5400 rpm or 500GB Serial ATA Drive @ 7200 rpm?
    Depends what you mean by "better." Faster? Yes. The 500 has 3x more storage capacity.
    Message was edited by: tjk

  • Questions about the differences between Arch and Chakra.

    I'm trying to decide whether to install Arch or Chakra on my laptop. Currently, I'm running Arch on my desktop and Windows 7 on my laptop (which I plan to overwrite). My skill level with GNU/Linux is somewhat intermediate – noobs refer to me as an expert, experts refer to me as a noob. If it matters, I'm a KDE user and primarily use my computers for web browsing and python development. Anyway, I just wanted to get an objective opinion on the differences between Arch and Chakra. I have a few points that stand out to me, but I welcome any input.
    -Stability
    I started my journey into GNU/Linux with Debian back in 2009 due to it's stability. I really disliked Debian's ancient software and considered moving my system to Unstable. While doing some googling about Debian Unstable, I stumbled across Arch – it was love at first sight. I've been an Arch user ever since. I love the bleeding edge software, and haven't had any major problems since I originally installed it. However, every time I run a system update I cringe a little. While Arch hasn't broken on me yet, I've read plenty of horror stories and it makes me uneasy. I understand that Chakra is a mix between a point and rolling release model. Is it any more or less stable than Arch? I know there are other distrobutions out there, but I'm in love with the Arch philosophy.
    -Security
    Pretty self-explanatory, but is there any difference in security between the two?
    -AUR
    As much as I love Arch, I wouldn't be able to stand it if it weren't for the massive collection of software available in the AUR. While I'm perfectly capable of compiling software myself, I prefer to use a command like tool like yaourt to manage my software. I understand that Chakra doesn't officially support the AUR and that they have their own user repository. Seeing as Chakra is still relatively new, is it lacking? Will I miss the AUR as a Chakra user?
    -Repositories
    Is there much difference in the official repositories between the two distrobutions?

    avonin wrote:
    I'm trying to decide whether to install Arch or Chakra on my laptop. ... I'm a KDE user ... I just wanted to get an objective opinion on the differences between Arch and Chakra..
    -Stability...
    -AUR...
    -Repositories...
    My take on Chakra is that it's the same as Arch with different developers.  They use pacman. They have a different and rather nice build system for their developers. They're doing a good job, but I'd hate to give up the services of Allan McRae who must work full time keeping the Archlinux core and toolchain up to date.  Chakra devs probably piggy-back off his work.
    As for "semi" rolling: I don't see Chakra as having a stable core.  A stable core sounds attractive, it would be like NetBSD which has a very stable core Unix operating system with apps added via pkgsrc.  But Chakra's core and toolchain is at the same version levels as Archlinux most of the time and are no more tested and stabilized than ours. Their core packages are updated piecemeal just like ours; there is no stable core that is released as a unit (afaik). Today Chakra has gcc 4.7 / glibc 2.15 just like ours. Their kernel is a little more stable: they're using udev 181 / linux 3.2.8 while Arch is on udev 182 / linux 3.3.7.  They are more conservative in upgrading xorg and the video drivers than Arch.  For example, today they're on xorg-server 1.10.4 / intel video 2.17 while Arch is up-to-the-bleeding-edge-minute with xorg-server 1.12.1.902 and intel video 2.19.  Yeah, I would consider Chakra to be a little more "stable" than Arch mainly because of their relaxed pace in changing the kernel and the xorg stuff.
    Most of the patches that I look at for Arch packages (I build my system entirely from source and try to build monthly releases for myself) are needed because we use more recent core packages like glib2/glibc/gcc than the developers of higher level stuff like qt.  Chakra is in the same situation.  We're on the front of the wave
    The Chakra CCR is compatible with the Arch AUR and mainly draws from AUR (an AUR buildscript will usually work fine on a Chakra system -- they just add one or two additional info fields.)  With a little effort you could get any package installed on a Chakra system that is available on Arch.
    Last edited by sitquietly (2012-05-24 20:43:58)

Maybe you are looking for

  • Printing from Reports

    Hi, I have HP Laser Jet 6L printer configured in Windows 98 machine and Windows 2000 machine. I have configured the Hp 6L drivers both in the client machines and linux servers. I made proper entry in uiprint.txt given by oracle for printing in Linux

  • Need help to start with some file and text manipulation

    Hello script mavens, I need help with starting a script that does the following: -within a base folder it takes an inventory (list?) of all the files (which happen to be image files). -creates a new folder inside the base folder and calls it imagesX

  • Is there any tutorials on how to make a form in Adobe Acrobat?

    http://www.bgsu.edu/downloads/hr/file26711.pdf I need help! I need to make a form just like the one in the link above. Is there any step by step directions on how to do this? Thanks! Salome Roe

  • Loading TXT documents in attachment list

    Hi, I have got a program that loads documents from a directory against the relevant PO so they appear in the attachment list when viewing the PO. My problem is for TXT documents when I try to open them they open with internal SAP editor rather than n

  • Can H323 gatekeeper in the same box as Gateway?

    The customer is goint to setup Polycom between 2 offices. With the limited budget, can I setup one 2610 working as both Gatekeeper and Gateway for each site? If yes, which IOS is recommended? Please provide a link for sample configuration. If no, whi