Best practices for accessing remote management

So, I've been looking into consolidating and moving our servers and such to a colocation datacenter. A problem for me that arises from moving is, what do we do about our remote access?In a private office enviornment, I haven't ever opened up VMWare vCenter to the open internet, nor have I ever opened up DRAC/iLO past our firewall to the net. I've always just had all that management stuff hanging out on its own subnet/VLAN and I haven't ever bothered with giving remote access to anyone, really. (Well, I did once set up a windows box to allow me to RDP in and opened up the firewall for that RDP so I could then access that management VLAN from that PC)Moving to a colocation facility makes me wonder, what does everyone else do for this? Would one have a VPN configured on a router in their colo space and remote in that way, and if the...
This topic first appeared in the Spiceworks Community

Harvard University recently announced that on June 19, 2015, it discovered an intrusion into the IT networks of the Faculty of Arts and Sciences and Central Administration."Since discovering this intrusion, Harvard has been working with external information security experts and federal law enforcement to investigate the incident, protect the information stored on our systems, and strengthen IT environments across the University," university provost Alan Garber and executive vice president Katie Lapp said in a statement."At this time, we have no indication that personal data, research data, or PIN System credentials have been exposed," Garber and Lapp added. "It is possible that Harvard login credentials (username and password) used to access individual computers and University email accounts have been exposed."...Read More
Read More

Similar Messages

  • IPS Tech Tips: IPS Best Practices with Cisco Remote Management Services

    Hi Folks -
    Another IPS Tech Tip coming up and this time we will be hearing from some past and current Cisco Remote Services members on their best practice suggestions. As always these are about 30 minutes of content and then Q&A - a low cost high reward event.
    Hope to see you there.
    -Robert
    Cisco invites you to attend a 30-45 minute Web seminar on IPS Best   Practices delivered via WebEx. This event requires registration.
    Topic: Cisco IPS Tech Tips - IPS Best Practices with Cisco Remote Management   Services
    Host: Robert Albach
    Date and Time:
    Wednesday, October 10, 2012 10:00 am, Central Daylight Time (Chicago,   GMT-05:00)
    To register for the online event
    1. Go to https://cisco.webex.com/ciscosales/onstage/g.php?d=203590900&t=a&EA=ralbach%40cisco.com&ET=28f4bc362d7a05aac60acf105143e2bb&ETR=fdb3148ab8c8762602ea8ded5f2e6300&RT=MiM3&p
    2. Click "Register".
    3. On the registration form, enter your information and then click   "Submit".
    Once the host approves your registration, you will receive a confirmation   email message with instructions on how to join the event.
    For assistance
    http://www.webex.com
    IMPORTANT NOTICE: This WebEx service includes a feature that allows audio and   any documents and other materials exchanged or viewed during the session to   be recorded. By joining this session, you automatically consent to such   recordings. If you do not consent to the recording, discuss your concerns   with the meeting host prior to the start of the recording or do not join the   session. Please note that any such recordings may be subject to discovery in   the event of litigation. If you wish to be excluded from these invitations   then please let me know!

    Hi Marvin, thanks for the quick reply.
    It appears that we don't have Anyconnect Essentials.
    Licensed features for this platform:
    Maximum Physical Interfaces       : Unlimited      perpetual
    Maximum VLANs                     : 100            perpetual
    Inside Hosts                      : Unlimited      perpetual
    Failover                          : Active/Active  perpetual
    VPN-DES                           : Enabled        perpetual
    VPN-3DES-AES                      : Enabled        perpetual
    Security Contexts                 : 2              perpetual
    GTP/GPRS                          : Disabled       perpetual
    AnyConnect Premium Peers          : 2              perpetual
    AnyConnect Essentials             : Disabled       perpetual
    Other VPN Peers                   : 250            perpetual
    Total VPN Peers                   : 250            perpetual
    Shared License                    : Disabled       perpetual
    AnyConnect for Mobile             : Disabled       perpetual
    AnyConnect for Cisco VPN Phone    : Disabled       perpetual
    Advanced Endpoint Assessment      : Disabled       perpetual
    UC Phone Proxy Sessions           : 2              perpetual
    Total UC Proxy Sessions           : 2              perpetual
    Botnet Traffic Filter             : Disabled       perpetual
    Intercompany Media Engine         : Disabled       perpetual
    This platform has an ASA 5510 Security Plus license.
    So then what does this mean for us VPN-wise? Is there any way we can set up multiple VPNs with this license?

  • Req:SAP Best practice for the Funds Management

    Dear all,
    Let me know where I can get the SAP Best practice for the Funds Management . Waiting for your valuable reply.
    Regards
    Manohar

    Hello Manohar,
    You can find documentation in links below:
    Industry Solution Master Guide - SAP for Public Sector:
    https://websmp105.sap-ag.de/~form/sapnet?_SHORTKEY=00200797470000065911
    SAP Best Practices for Public Sector:
    http://help.sap.com/  SAP Best Practices -> Industry Packages -> Public
    Sector
    Online Library for Funds Management:
    http://help.sap.com/saphelp_erp2005vp/helpdata/en/41/c62c6d6d84104ab938a
    a7eae51db06/frameset.htm
    I hope it helps
    Best Regards,
    Vanessa Barth.

  • Best Practices for Accessing the Configuration data Modelled as XML File in

    Hi,
    I refer the couple of blof posts/Forum threads on How to model and access the Configuration data as XML inside OSB.
    One of the easiest and way is to
    Re: OSB: What is best practice for reading configuration information
    Another could be
    Uploading XML data as .xq file (Creating .xq file copy paste all the Configuration as XML )
    I need expert answers for following.
    1] I have .xsd file which is representing the Configuration data. Structure of XSD is
    <FrameworkConfig>
    <Config type="common" key="someKey">proprtyvalue</Config>
    <FrameworkConfig>
    2] As my project will move from one env to another the property-value will change according to the Environment...
    For Dev:
    <FrameworkConfig>
    <Config type="common" key="someKey">proprtyvalue_Dev</Config>
    <FrameworkConfig>
    For Stage :
    <FrameworkConfig>
    <Config type="common" key="someKey">proprtyvalue_Stage</Config>
    <FrameworkConfig>
    3] Let say I create the following Folder structure to store the Configuration file specific for dev/stage/prod instance
    OSB Project Folder
    |
    |---Dev
    |
    |--Dev_Config_file.xml
    |
    |---Stage
    |
    |--Stahe_Config_file.xml
    |
    |---Prod
    |
    |-Prod_Config_file.xml
    4] I need a way to load these property file as xml element/variable inside OSb message flow.?? I can't use XPath function fn:doc("URL") coz I don't know exact path of XMl on deployed server.
    5] Also I need to lookup/model the value which will specify the current server type(Dev/Stage/prod) on which OSB MF is running. Let say any construct which will act as a Global configuration and can be acccessible inside the OSb message flow. If I get the vaalue for the Global variable as Dev means I will load the xml config file under the Dev Directory @runtime containing key value pair for Dev environment.
    6] This Re: OSB: What is best practice for reading configuration information
    suggest the designing of the web application which will serve the xml file over the http protocol and getting the contents into variable (which in turn can be used in OSB message flow). Can we address this problem without creating the extra Project and adding the Dependencies? I read configuration file approach too..but the sample configuration file doesn't show entry of .xml file as resources
    Hope I am clear...I really appreciate your comments and suggestion..
    Sushil
    Edited by: Sushil Deshpande on Jan 24, 2011 10:56 AM

    If you can enforce some sort of naming convention for the transport endpoint for this proxy service across the environments, where the environment name is part of the endpoint you may able to retrieve it from $inbound in the message pipeline.
    eg. http://osb_host/service/prod/service1 ==> Prod and http://osb_host/service/prod/service2 ==> stage , then i think $inbound/ctx:transport/ctx:uri can give you /service/prod/service1 or /service/stage/service1 and applying appropriate xpath functions you will be able to extract the environment name.
    Chk this link for details on $inbound/ctx:transport : http://download.oracle.com/docs/cd/E13159_01/osb/docs10gr3/userguide/context.html#wp1080822

  • Best practice for using remote control under limited rights?

    Hi. We are getting ready to take admin rights away from our users and make them standard users. We plan to utilize Zen for most of our in-scope applications so that we can allow users to install supported software. There is usually no problem in that case because Zen can elevate to System access during the install. However, we know that there are applications out there that a user may want to install that is not packaged in Zen. Also, in the event that a system setting needs to be changed, we will have to have a method for supporting this. In either case, the user will call our help desk. Unfortunately, the user will not have enough rights to do the install or system change even if the help desk associate remote control's the PC. What is the best practice to handle this situation in a Netware/Zenworks environment where users only have limited access?
    I was thinking of three possibilities:
    1.) The obivous one is to send a technician over to log in using local admin credentials to install the software or perform the change. (Drawback - not very efficient because a desktop tech would have to get over to the user's PC to perform the work)
    2.) Have the help desk engineer log out of the machine through remote control, log back in as local admin to install the software or perform the change. (Drawback - not very convienant and time consuming.)
    3.) Have the help desk engineer use the "run as" command or even create a Zen application object that could be executed to provide temporary rights for installing software or making system changes. Aaron Margosis of Microsoft writes about this quite a bit in his blog Aaron Margosis' "Non-Admin" WebLog : Table of Contents (Aaron Margosis' Non-Admin WebLog) (Drawback - some software or settings will not work properly using this technique)
    The last one that I didn't list was creating a new application object. I did not factor this one in because this isn't always applicable to system changes and we really don't want to be making app objects for every out of scope app that exists in the user community. We typically only make them for widely used and supported apps.
    Your feedback is appreciated.
    Thanks

    Originally Posted by spond
    Joshbilsky,
    how about
    4) use the remote execute option to remotely launch an app as admin?
    Shaun Pond
    That's probably an option that we will make available. I wasn't sure how some things will work under the SYSTEM context vs local admin.

  • SAP HANA Security - Best Practice for Access to Schemas??

    Hi,
    Currently we don'y have a defined Security model in HANA Studio.Neither there is no defined duties of a BASIS / Security / Developers.
    I want to understand what best practices are followed at other customers for defining security for Schema.
    1. Who should be creating the schema for Developers / Modelers?
    2. Should we use our own ID's to create/maintain these Schema or a Generic ID?
    Right now, when developers log in to Studio, by default they are assigned to their own schema (User ID) and they create objects under that.
    We(Security team), face issues when other developers need access to schema of another user as they want to develop objects under schema of different user
    Also, who should be owning the "SYSTEM" user ID and what steps needs to be done whenever a new schema is created.
    Thanks for the help in advance.

    Hi,
    I created a project (JDeveloper) with local xsd-files and tried to delete and recreate them in the structure pane with references to a version on the application server. After reopening the project I deployed it successfully to the bpel server. The process is working fine, but in the structure pane there is no information about any of the xsds anymore and the payload in the variables there is an exception (problem building schema).
    How does bpel know where to look for the xsd-files and how does the mapping still work?
    This cannot be the way to do it correctly. Do I have a chance to rework an existing project or do I have to rebuild it from scratch in order to have all the references right?
    Thanks for any clue.
    Bette

  • Best practice for limiting network management to few devices

    Hello ,
    I have set up a very basic security implementation that is no way realistic, but I just want to experiment and learn...
    In my 1801 router that answers DHCP requests on separate wired and wireless vlans, I have bound static IP addresses to the MAC addresses of my laptop wireless and wired interfaces.
    Then I set up an ACL to permit inbound traffic from these IPs only for the vty lines.
    Obviously this is easily defeated by statically assigning these same IPs to any device on the network, so I was thinking about a better way to limit management of the router to a few devices.
    What is the best practice in professional environments?
    Thanks.

    Obviously this is easily defeated by statically assigning these same IPs to any device on the network, so I was thinking about a better way to limit management of the router to a few devices.
    TACACs or RADIUS with robust password policy and regular interval to change the passwords (30 to 45 days).
    Read this and go to the "Composing hard-to-guess passwords" section.

  • Best practice for highly available management / publishing servers

    I am testing a highly available appv 5.0 environment, which will deploy appv packages to a Xenapp farm.  I have two SQL 2012 servers configured as an availability group for the backend, and two publishing / management servers for the front end. 
    What is the best practice to configure the publishing / management servers for high availability?  Should I configure them as an NLB cluster, which I have tested and does seem to work, or should I just use the GPO to configure the clients to use both
    publishing servers, which I have also tested and appears to work?
    Thanks,
    Patrick Sullivan

    In App-V 5.0 the Management and Publishing Servers are hosted in IIS, so use the same approach for HA as you would any web application.
    If NLB is all that's available to you, then use that; otherwise I would recommend a proper load balancing solution such as Citrix NetScaler or KEMP LoadManager.
    Please remember to click "Mark as Answer" or "Vote as Helpful" on the post that answers your question (or click "Unmark as Answer" if a marked post does not actually
    answer your question). This can be beneficial to other community members reading the thread.
    This forum post is my own opinion and does not necessarily reflect the opinion or view of my employer, Microsoft, its employees, or other MVPs.
    Twitter:
    @stealthpuppy | Blog:
    stealthpuppy.com |
    The Definitive Guide to Delivering Microsoft Office with App-V

  • Best Practices for many Remote Objects?

    Large Object Model doing JDBC over RMI
    Please read the following and provide suggestions for a recommended approach or literature that covers this area well?
    N-Tiered Architecture
    JSP/Servlet (MVC) - Database Access Layer - Database
    Applets - JSp/Servlet Engine - Database Access Layer - Database
    Application Layer - Application Layer - Database Access Layer - Database
    I have an object model developed using Torque (A JDBC Object Relational modeling framework) for over 100 tables (to be over 160) that I am commencing to enable over RMI. I have got several remote methods up and running. For some of the simple methods starting up has been easy. Going forward I forsee issues.
    Each table has a wrapper or data object and a peer object that have setters/getters and special methods as desired. The majority of these classes are extended from Base Objects that have basic common functionality for retrieving, creating, and manipulating with a database using SQL.
    I have started building a Remote Interface and an Implementation class that invoke the necessary methods and classes within the Object Model to pull successfully off or update the database. Additionally the methods will need to return objects that represent non-primitive serializable dataobjects and collections of objects.
    Going forward client applications, servlets, and jsps will be using the database in more complex and comprehensive methods over rmi. Here are a couple of things I am concerned about.
    1) When to use java.rmi.server.codebase for class loading? In my implementation several of the remote methods will return objects (e.g Party, Country, CountryList, AccountList). These objects themselves are composed of other objects that are not part of the jvm. For all remote methods that return non-primitive objects must you include the classes in the codebase for the client to operate upon them. Couldn't this be pointless as you have abstract and extended classes all residing within the codebase? In practice do people generally build very thin proxy objects for the peer/data objects to hold just the basic table elements and sets?
    2) Server Versioning/Identity - Going forward more server classes will be enabled via rmi. Everytime one wants to include more methods on interface that is available must you update the interface, create new implementation classes, and redistribute the Remote Interface, and stubs/skels to client apps to operate on again? Is there some sort of list lookup that a client can do to say which processes are available remotely presently (not just at initialization)? As time changes more or fewer methods might be available?
    Any help is greatly appreciated.

    More on Why other approaches would be better?
    I have implemented some proxy objects for the remote Data Objects produced by Torque. To ease the pain, I have also constructed a proxy Builder that takes the table schema and builds a Proxy Object, an inteface for the proxy, and methods to copy between the Torque Data Object (which only lives on the server) into the Proxy Object (accessible by client and/or server).
    The generated methods are useable in the object implementing the Remote Interface but are themselves not remoteable. Only the Server would use these methods. Clients can only receive primitives, proxy Objects, or collections of ProxyObjects.
    This seems to be fairly light currently. I had to jump hoops to use Torque and enable remote apps to use the proxy objects. What would be the scaling issues that will come up? Why would EJBs with containers and all kinds of things about such as CMP vs BMP to be concerned with be a better approach?
    Methods can be updated to do several operations verses Torque and return appropriately (transactions). In this implementation the client (Servlet, mini App or App) needs the remote stub and the proxy objects (100 or so) to stand in for the Torque generated Data Objects. A much smaller and lighter set of classes, based on common JDK classes, instead of the torque classes (and necessary abstracts/objects/interfaces/exceptions for Torque).

  • Best Practices for Organizational Change Management

    Hooray!  Finally an opportunity to take this forum back to its proper direction of Organizational Change Management topics as we begin to engage with some subject matter experts in OCM.
    Those interested in having OCM conversations here are also welcome to begin interaction with this <a href="https://https://www.benchmarking.sap.com/cgi-bin/qwebcorporate.dll?idx=945JDN&SHSP1Q2A=asugEMAIL061907">OCM best practices survey</a>.  It would give community a chance to engage with ASUGers around the topic of organizational change and give ASUGers a chance to broaden the Americas centric perspective.  So welcome all.
    Looks like
    <a href="/people/kerry.brown/blog/2007/07/10/organizational-change-management-best-practices-survey Brown</a> will be helping put this topic back on course.  Looking at her blog profile, I, for one, am most eager for her engagement with us here.

    Hai,
    Organizational Structure
    Use
    In order for the workflow system to establish the relationship between the requester and their superior, you must create an organizational plan in the system.
    For this tutorial, of course, this organizational plan does not have to be complete and valid across the enterprise. To keep the test as simple as possible initially, define an organizational plan that only contains one administrator and one head of department.
    Assign both items to yourself. As a result, all work items will appear in your own Business Workplace. Later you will modify the organizational plan such that you will have to work through the scenario with two users.
    Procedure
    The organizational plan required for this tutorial consists, when complete, of one organizational unit (= "department"), which contains two positions: a head of department and an administrator.
    Each position is described by one job and each position is assigned one user as holder. The head of department position is also designated as chief position of the organizational unit.
    Of course, a "real" organizational plan is created by arranging several organizational units with their positions in a hierarchy. Usually several positions are described by one job.
    For further information, refer to the documentation  Organizational Plan.
    The procedure in this unit is divided into four parts:
    1.     You create an organizational unit.
    2.     Enter necessary jobs as required.
    3.     You create a position for the head of department in the new organizational unit.
    You define a position in three steps:
    i.     You create a position that is assigned your organizational unit.
    ii.     You assign a holder to the position.
    iii.     You assign a job to the position.
    3.     You create a position for the administrator in the organizational unit.
    Creating an organizational unit
    1.     Choose Tools &#61614; SAP Business Workflow &#61614; Development &#61614;&#61472;Definition tools &#61614; Organizational Management &#61614; Organizational plan &#61614; Create.
    2.     Confirm the validity period proposed in the dialog box Creating a Root Organizational Unit.
    This takes you to the Create Organization and Staffing (Workflow) screen. This user interface is divided into four screen areas:
    3.     On the Basic Data tab in the details area, enter an abbreviation and a name in the Organizational unit input fields.
    Abbreviation: <ini_sales>
    Name: <OrgUnit: Sales (ini)>
    4.     Choose  .
    You can now create the position for a head of department and one administrator.
    Create jobs
    When enhancing an organizational unit, the necessary jobs are usually already available. For this tutorial however, you create the necessary jobs for head of department and administrator yourself.
    1.     Choose Edit &#61614; Create jobs.
    You go to the dialog box Create jobs. The lower area contains a list of existing jobs and the upper area contains an input table in which you can create new jobs by entering abbreviations and names.
    2.     In the input table, enter an abbreviation and a name for each of the new jobs.
    Job - head of department:
    Object abbreviation: < ini_dhead_C >
    Name: < job: head of department (ini) >
    Job administrator:
    Object abbreviation: < ini_admi_C >
    Name: < job: administrator (ini) >
    3.     Choose  .
    Creating Position for Head of Department
    1.     Change to the overview area in the staff assignments of the organizational unit, in order to assign positions, jobs and holders. Choose the arrow  on the right next to the  and then the staff assignments (list).
    2.     Choose  .
    A new position is then created in the staff assignments and is displayed in a new line in the table. The position is vacant and no job is assigned to it.
    3.     Open the details view for the new position in the details area by double-clicking on the entry in the table.
    4.     On the Basic Data tab, enter a code and a description in the Position input fields. Overwrite the previous contents.
    Abbreviation: < ini_dhead_S >
    Description: < position: head of department ( ini ) >
    Assigning a holder to the position
    You now assign R/3 users to the positions. The staff assignments for your organizational unit are displayed and you see the vacant position in the table.
    5.     Select User in the search area and enter the search criteria in order to find your user names.
    All of the user names that match your search criterion are listed in the selection screen.
    6.     Select your user name in the selection area and drag it to the Person/User column of the position in the overview area.
    Confirm the message that the relationship period of the validity has been changed.
    7.     Set the Head of own organizational unit indicator in the details area.
    Assigning a job to the position
    Assign the job of the head of department you created earlier to the position.
    8.     Select Job in the search area and enter the search criteria in order to find the job of the head of department.
    All jobs that match your search criterion are listed in the selection screen.
    9.     Select job: head of department ( ini) in the selection area and drag it to the Job column of the position in the overview area.
    10.     Choose  .
    The job is assigned the position. Check this by switching to the staff assignments of the organizational unit. Select the organizational unit in the overview area, choose the arrow  on the right next to the  and then the staff assignments (list). The newly created job is displayed in the job column.
    Create position for administrator and assign holder and job
    You are now in the staff assignments of the organizational unit.
    1.     Choose  .
    A new position is then created in the staff assignments and is displayed in a new line in the table. The position is vacant and no job is assigned to it.
    2.     On the Basic Data tab in the details area, enter an abbreviation and a name in the Position input fields. Overwrite the previous contents.
    ID: <ini admiS >
    Description: < position: administrator ( ini ) >
    3.     Choose  .
    4.     Select User in the search area and enter the search criteria in order to find your user names.
    5.     Select your user name in the selection area and drag it to the Person/User column of the position in the overview area.
    Confirm the message that the relationship period of the validity has been changed.
    6.     Select Job in the search area and enter the search criteria in order to find the job of the administrator.
    7.     Select job: administrator ( ini) in the selection area and drag it to the Job column of the position in the overview area.
    8.     Choose  .
    Result
    Display your entire staffing schedule again and make sure that all the information listed is correct.
    You can display a detailed view of jobs, users, and positions. Choose the relevant cell in the table by double-clicking it.
    In the details view of a position or job, all of the assigned tasks are displayed on the Tasks tab.
    You have now completed the first unit (defining the organizational plan). You can now start on the next unit. To exit processing of the organizational structure, choose Back.

  • Best Practice for VOD Remote Storage

    Hi All,
    We have a couple of servers that host FLV videos for on-demand playback.  The file storage however is on a separate server, and all the stored VOD content is accessed via UNC share, ie \\servername\share\foo\bar.  The servers have a fast connection to the file server, however, whenever we try to play videos from the FMS servers, the first time its requested, they play back jerky and jumpy.  Most of the time though, the second time the video is played, it plays fine.  Also, if you try to seek ahead to a spot where no one's played back from before, it will be jerky the first time, but fine the second time.
    I'm assuming that the first time through, the FMS server caches the content, and then the 2nd time it servers it from the cache.  Since the servers have a very fast connection between them, what can be done to solve this?
    Also, would it be better to have an Origin server host the file content locally, and then use Edge servers to connect to and serve the files?  Would this alleviate the problem at all?  Thank you for any help.
    Best,
    -Costin

    We're probably going to have to define 'fast connection' a bit here.  It sounds like either the speed of requests needed for FMS, or their size is causing an issue getting sufficient data flow to FMS to avoid stuttering on the desired file.  There's a few ways that you can measure this.  You can go the route of measuring the I/O on network from your UNC server location, or if you're running an FMIS file adaptor you can plug in and log the speed of return of your reads.  At the end of it though it's going to be about the numbers of how quickly we're able to answer FMS' requests for data and demonstrating that we do or don't have enough flowing through here.
    Asa

  • Best practice for core data managed objects

    Hello
    I'd like to konw if there is a document available listing the good practices when managing core data managed objects.
    For example should I keep those objects in memory in a singleton class, or save thme to the DB and load them when needed, ... I am trying to figure out how to manage Annotation views representing managed objects when using the MapKit.
    Thanks

    Seen this?
    Using Managed Objects

  • "Best practice" for accessing a class from a custom component?

    My app utilizes a simple class to hold global properties such as username, session data, and similar data. The class is initialized at app startup via code similar to: appGlobals:myGlobals=new myGlobals.
    Many of the custom MXML components and AS classes need to access that data. I have been able to work with it using Application.application.appGlobals.propertyname.
    Is this method the best way to communicate from components and classes to a class initiated at the application level, or should I learn something new before I build a lot of code on this method?
    Thanks.
    Paul

    The WizardModel class is interesting, it is a "singleton" where it is designed to only have one instance, and the class actually has a static variable of its own class. Because that variable is static, an instance is created the first time the class is accessed.
    As to where the WizardModel is "first accessed" and thus its own variable of type WizardModel instantiated, is hard to say, as you really need to understand the application and component startup lifecycle indepth. I have a certain depth of knowledge of that but not enough depth to say definitively when WizardModel  is first accessed, but here are some possibilities:
    WizardModel.wizardTitle = WizardModel.wizardTitleBase;      In the WizardController "wizardTitleChangeHandler" event handler
    creationComplete="WizardModel.app = this;"      In the Wizard.mxml main app creationComplete handler
    <mx:Panel title="{WizardModel.wizardTitle}" width="100%" height="100%">    Opening tag of Panel in Wizard.mxml
    I know its confusing, but just try to absorb what you can for now, and over time it will become gradually more clear.

  • Best practices for accessing data in subviews

    I've got two iPhone projects which share most of their code base. I'm trying to figure out the best way to load data from some plist files and store them in a common container UIView and provide access to the data for subviews and subviews of the subviews, etc. Right now I've got the data being passed from the container view to the subviews it creates and then the subview themselves pass it further, basically a bucket brigade to get the data to where it needs to go which could be 3 or 4 views down in the hierarchy.
    Is there a better approach? I've looked at delegates & protocols but I'm having a hard time understanding how they work and whether they are appropriate in this situation. Originally I had the app delegate holding the data and any class anywhere could invoke the app delegate and get the data. However this approach fails with 2 projects because the app delegates have different names and the classes that need to access it are common to both projects. Can the app delegate be renamed without significant impact? Or is there a way a UIView can be set up as a delegate in much the same way?
    Thanks for any advice!
    Greg

    Hi Greg - You can rename the app delegate class to whatever you want, just remember to change it in IB, and make sure you catch any place it already appears in your code. I guess there's no need to change the app delegate class file names unless you want to.
    However there are lots of other solutions to your problem. A case could be made for declaring a global pointer to this data, for example. Or, you could encapsulate the data wherever you want and make an extern (globally visible) C function to access it.
    Another solution would be to put the data in a shared object which would be accessed just like the shared app object, e.g.
    #include "MyObject.h"
    NSDictionary *myPlistData = [MyObject sharedObject].plistData;
    I just got done looking at "how to make a shared object" in the Cocoa docs and can't seem to find it atm. Anyway it's in there somewhere, either in an Obj-C doc or one of the top level guides. The job just wants a class method that returns the pointer stored in a static C var; if the var is nil, the object is first created and its addy is stored in the var.
    Hope that helps!
    - Ray

  • What is best practice for remotely managing bank of switches over POTS

    I need to be able to have a back door into several catalyst switches and ASA.
    What is the best practice for accessing them remotely. ?

    Just place a modem into any console port. Ideally you use a terminal server, but is not always really needed.

Maybe you are looking for

  • Problems "pulling" photos from iPhoto into iMovie....

    I am trying to create a movie with a combination of movie clips from my DV camera and photos from iPhoto (on my G5, 2ghz iMac). Movie clips are no problem. The problem occurs when I try and "pull" in more than 20 pics or so from iPhoto. I am trying t

  • Query to see all the tables in db

    I want all tables seeing in the PLSQL

  • I Use Private Browsing Always, and Still Get Sanitize.js:136 Error

    All add-ons, extensions, plug-ins, and Firefox itself (v. 7.0.1) are up-to-date. I ALWAYS use Private Browsing, but still get the chrome://browser/content/sanitize.js:136 error. After "exiting" Firefox, Firefox refuses to close right away. I either w

  • Strange behaviour in web cartographic application

    Hi group, I have noticed a strange behaviour with the overview tool of my web application with MapInfo data uploaded in Oracle 10g with EasyLoader. When I put the cursor near the limit of some polygons but outside of its surface, the overview tool sh

  • How to turn off service led in a SF 6900

    Hello I have a SF6900 (the big server) and the service yellow light on top of the rack is on. I can see it on with showplatform on the SC. There are no HW faults in this server. Does anyone know how to turn of this led? I copied below part of the sho