What components do we need to implement NAP solution ?

We use win2003R2 and win2008R2 servers , win7 laptop,win2008 DHCP, cisco 802.11x switch,Access Point, cisco acs radius server, windows CA.
I am planing to implement NAP solution.
What components do we need to implement NAP solution ?
I think it needs NAP server role and third-party antivirus Win NAP compatible software.

Hi,
As we know, there are five NAP enforcement methods, NAP Enforcement for IPSec Communications, NAP  Enforcement for 802.1x, NAP Enforcement for DHCP, NAP Enforcement for Terminal Services Gateway, NAP Enforcement for VPN.
Due to you have many devices in your environment, which is the primary means that you want to deploy NAP? In my point of view, you could configure NAP enforcement for DHCP or for 802.1x Wireless or for 802.1x Wired.
To deploy DHCP NAP, we need a DC, a NPS server as RADIUS server, a DHCP server as a RADIUS clients and a DHCP client.
To deploy NAP Enforcement for 802.1x Wireless, we need a DC, a CA, a NPS server as RADIUS server, a 802.1x wireless access point as a RADIUS client and a 802.1x wireless client.
To deploy NAP Enforcement for 802.1x Wired, we need a DC, a CA, a NPS server as RADIUS server, a 802.1x authenticating switch as a RADIUS client and a 802.1x wired client.
You could choose one of these NAP enforcement based on your environment. More details about NAP configuration, please refer to links below,
Checklist: Configure NAP Enforcement for DHCP
http://technet.microsoft.com/en-us/library/cc772356(v=ws.10).aspx
Checklist: Configure NAP Enforcement for 802.1X Wireless
http://technet.microsoft.com/en-us/library/cc753793(v=ws.10).aspx
Checklist: Configure NAP Enforcement for 802.1X Wired
http://technet.microsoft.com/en-us/library/cc730926(v=WS.10).aspx
Best Regards,
Tina

Similar Messages

  • We need to implement EP Solution for Supplier Portal

    Hello All,
         I'm working on BSP's and Portal since 1 Year.
    The Portal Implementation is already done before I started working on it.
    Now my scenario is as folows :
    We need to implement EP Solution for Supplier Portal .
    For this we want to know what are all the activities involved.
    Can anyone throw some light on it ?
    Regards,
    Deepu.K

    We have to make setting for displaying all the roles defined for supplier portal to reflect in the EP.Portal is integrated with Webclient of CRM so make sure all changes done in the webclient are reflected in the Portal thing.

  • What components do I need to develop and run JSP applications?

    I am currently developing a JSP-based application using Apache, Tomcat, MySQL, and JDBC. I'm considering switching to Oracle since I know it better and it has better support for certain capabilities.
    But, it's very confusing trying to determine from the Oracle sites just what I need and which platforms are supported. What do I need to buy and what can I download?
    So, which Oracle components do I need to develop and test JSP applications using Application Server? I'll probably get into XML as well.
    Does any of it run on Win98, WinNT, Win2000, or SolarisIntel? I likely will not deploy on any of these platforms, but it would be convenient to develop on my Intel laptop, if possible.
    --Terry Westley, [email protected]                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                   

    Oracle's IAS includes an Apache and Jserv with JSP runtime by default.
    You could simply migrate your Apache/Tomcat JSP to Apache/Jserv/Oracle JSP without problems, including win32 plattform.
    To run Servlet compliant with JSDK 2.2+ you has to use Apache mod_ose module with your servlets and jsp running inside the Oracle JVM, this combination is more scallable and secure for big projects, but it requires more hardware for your laptop.
    Best regards, Marcelo.

  • What components do I need to run my CR XI reports from another application?

    Post Author: rickcf
    CA Forum: Data Connectivity and SQL
    We have just started using CR XI and have written 2 reports that is parameter driven and run off an Oracle database.  We have an application written in Java that also runs off this database.  We want to have the java application run the two reports by passing in the parameters to the reports.  Can anyone tell me how I can accomplish this from a CR stand point.  Any components I might need install or have running to make CR XI run nice with the Java client/server application.  Any things we will need on the workstation side to make the reports run for all the users.  Any info or insight appreciated. 
    Thanks

    Post Author: synapsevampire
    CA Forum: Data Connectivity and SQL
    Here are the Java sample apps:
    http://support.businessobjects.com/communityCS/FilesAndUpdates/sample_applications_for_Java_developers.pdf.asp
    -k

  • What components do I need to get?

    Im getting totally confused with all these different packages.
    I want to create a flex app that connects to mysql.  I downloaded and installed the flex 3 plugin, tomcat, and the tomcat plugin.
    So now I can create a simple little app and show it from the server.
    But when I create a new flex project, it asks for server technology, which I guess is J2EE.  What is the remote object access service?  And lifecycle data services?
    Because if I select that I need to then select a flex war file, which I have no idea about.  Where is it?  Do I need to get it?
    And what about BlazeDS?  Do I need that?
    Or the lifecycle data service package?  Is that part of the flex plugin I installed (since I can select it with new flex project)?  And that install packages comes with its OWN tomcat.  I like to have tomcat run via eclipse...
    What do I need to be using here, and what do I not need to worry about?

    LCDS stands for LiveCycle Data Services:
    http://www.adobe.com/products/livecycle/dataservices/
    http://www.adobe.com/products/livecycle/dataservices/indepth.html
    In a nutshell, LCDS allows realtime data push/pull - keeping applications synchronized.
    BlazeDS is the open source variant:
    http://opensource.adobe.com/wiki/display/blazeds/BlazeDS/
    How to set up a BlazeDS project:
    http://learn.adobe.com/wiki/display/Flex/Creating+Flex+Builder+Projects+that+Use+Server+Te chnologies
    If you want a Flex application to "talk" to your server (e.g. to load data from MySQL) you can use Flex Remoting, which is a binary messaging format (AMF3). To do so, you'll need "something" on your server that understands that format, LCDS/BlazeDS is that "something".
    Remoting comes in different flavors:
    PHP - AMFPHP, ZendAMF
    ColdFusion - built in
    J2EE - BlazeDS, LCDS
    etc..
    In Flex you use <mx:RemoteObject /> for Flex Remoting:
    http://livedocs.adobe.com/flex/3/langref/mx/rpc/remoting/mxml/package-detail.html
    Mind you, Remoting is not the only way to communicate with a server.
    You can send and load data (GET/POST) using <mx:HTTPService /> and use soap services with <mx:WebService />
    http://livedocs.adobe.com/flex/3/langref/mx/rpc/http/mxml/HTTPService.html
    http://livedocs.adobe.com/flex/3/langref/mx/rpc/soap/mxml/package-detail.html

  • SWs needed to implement SSO using AM

    Hi,
    I want to know what all SWs are needed to implement SSO using sun AM.
    I tried implementing SSO using AM along with Policy Agent. But somewhere I am missing something.
    Does anyone know of any simple doc which explains the steps in clearly with confusion?
    Thanks
    Rahul A Honrao

    Try these links, good start ups:
    http://www.sun.com/software/products/access_mgr/index.jsp
    http://developers.sun.com/identity/reference/techart/install.html
    http://docs.sun.com/app/docs/prod/sjs.policy.agt22~1322.1#hic

  • In what order what components needed to be configured in ODI?

    I will be migrating data from TXT files to Oracle 10g.
    In what order what components/objects (e.g. Project, Models, topology, data store, metadata, schema etc) needed to be configured in ODI?
    This will help me go through ODI manual in a logical order.

    Hi,
    Have a look as a kick off,
    http://www.oracle.com/technology/obe/fusion_middleware/odi/ODIproject_flatfile-to-table/ODIproject_flatfile-to-table.htm
    Thanks,
    Guru

  • What are the steps (High level) needed to implement US Payroll?

    Hi,
    What are the (high level) steps needed to implement US Payroll?
    Please provide any relavent URLs for any steps if possible.
    Thanks,
    Prabakar

    Transaction Code:-OH00
    Personnel Management Personnel AdministrationOrganizational Data Organizational Assignment Define employee attributes
    Transaction Code:-OH00
    Personnel Management Personnel AdministrationOrganizational Data Organizational Assignment Define employee attributes
    Transaction Code:-OH00
    Personnel Management Personnel AdministrationOrganizational Data Organizational Assignment Define employee attributes
    Transaction Code:-OH00
    Personnel Management Personnel AdministrationOrganizational Data Organizational Assignment Create payroll area
    Transaction Code:-OH00
    Personnel Management Personnel AdministrationOrganizational Data Organizational Assignment Create payroll area
    Transaction Code:-PE03
    Personnel Management Personnel AdministrationOrganizational Data Organizational Assignment Check Default Payroll Area
    Transaction Code:-PA03
    Personnel Management Personnel AdministrationOrganizational Data Organizational Assignment Create control record
    Transaction Code:-PA03
    Personnel Management Personnel AdministrationOrganizational Data Organizational Assignment Create control record
    Transaction Code:-OH00
    Personnel Management Personnel AdministrationPayroll DataBasic PayDefine EE Sub Group Grourping for PCR and Collective Agreement Provision
    Transaction Code:-OH00
    Personnel Management Personnel AdministrationPayroll DataBasic PayDefine Reason for Change
    Transaction Code:-OH00
    Personnel Management Personnel AdministrationPayroll DataBasic PayCheck PayScale Type
    Transaction Code:-OH00
    Personnel Management Personnel AdministrationPayroll DataBasic PayCheck PayScale Area
    Transaction Code:-OH00
    Personnel Management Personnel AdministrationPayroll DataBasic PayCheck Assignment of PayScale Structure to Enterprise Structure
    Transaction Code:-OH00
    Personnel Management Personnel AdministrationPayroll DataBasic PayDetermine Default for PayScale Data
    Transaction Code:-OH00
    Personnel Management Personnel AdministrationPayroll DataBasic PaySetup Payroll Period for Collective Agreement Provision
    Transaction Code:-OH00
    Personnel Management Personnel AdministrationPayroll DataBasic PayDefine PayScale Salary ranges
    Transaction Code:-OH11
    Personnel Management Personnel AdministrationPayroll DataBasic PayWage Types Create Wage Type
    Transaction Code:-OH11
    Personnel Management Personnel AdministrationPayroll DataBasic PayWage Types Create Wage Type
    Transaction Code:-OH00
    Personnel Management Personnel AdministrationPayroll DataBasic PayWage Types Check Wage Type Group u201CBasic Payu201D
    Personnel Management Personnel AdministrationPayroll DataBasic PayWage Types Check Wage Type CatalogCheck Wage Type Text
    Transaction Code:-OH13
    Personnel Management Personnel AdministrationPayroll DataBasic PayWage Types Check Wage Type CatalogCheck Entry Permissibility Per Infotype
    Transaction Code:-OH13
    Personnel Management Personnel AdministrationPayroll DataBasic PayWage Types Check Wage Type CatalogCheck Wage Type Characteristics
    Transaction Code:-OH13
    Personnel Management Personnel AdministrationPayroll DataBasic PayWage Types Check Wage Type CatalogCheck Wage Type Characteristics
    Transaction Code:-OH00
    Personnel Management Personnel AdministrationPayroll DataBasic PayWage Types Employee Sub Group Grouping for Primary Wage
    Transaction Code:-OH00
    Personnel Management Personnel AdministrationPayroll DataBasic PayWage Types Personnel Sub Area Grouping for Primary Wage Type
    Transaction Code:-OH00
    Personnel Management Personnel AdministrationPayroll DataBasic PayWage Types Define Wage Type Permissibility for each PS and ESG
    Transaction Code:-OH00
    Personnel Management Personnel AdministrationPayroll DataBasic PayWage Types Define Wage Type Permissibility for each PS and ESG
    Transaction Code:-OH00
    Personnel Management Personnel AdministrationPayroll DataRecurring Payment and DeductionDefine Reason for Change
    Transaction Code:-OH11
    Personnel Management Personnel AdministrationPayroll DataRecurring Payment and DeductionWage TypesCreate Wage Type Catalog
    Transaction Code:-OH11
    Personnel Management Personnel AdministrationPayroll DataRecurring Payment and DeductionWage TypesCreate Wage Type Catalog
    Transaction Code:-OH00
    Personnel Management Personnel AdministrationPayroll DataRecurring Payment and DeductionWage TypesCheck Wage Type Group u201C Recurring Payments and Deductionu201D
    Transaction Code:-OH13
    Personnel Management Personnel AdministrationPayroll DataRecurring Payment and DeductionWage TypesCheck Wage Type CatalogCheck Wage Type text
    Transaction Code:-OH13
    Personnel Management Personnel AdministrationPayroll DataRecurring Payment and DeductionWage TypesCheck Wage Type CatalogCheck Entry Permissibility Per Infotype
    Transaction Code:-OH13
    Personnel Management Personnel AdministrationPayroll DataRecurring Payment and DeductionWage TypesCheck Wage Type CatalogCheck Wage Type Characteristics
    Transaction Code:-OH13
    Personnel Management Personnel AdministrationPayroll DataRecurring Payment and DeductionWage TypesCheck Wage Type CatalogCheck Wage Type Characteristics
    Transaction Code:-OH00
    Personnel Management Personnel AdministrationPayroll DataRecurring Payment and DeductionWage TypesDefine Employee Sub Group Grouping for Primary Wage Type.
    Transaction Code:-OH00
    Personnel Management Personnel AdministrationPayroll DataRecurring Payment and DeductionWage TypesDefine Personnel Area Grouping for Primary Wage Type
    Transaction Code:-OH00
    Personnel Management Personnel AdministrationPayroll DataRecurring Payment and DeductionWage TypesDefine Wage Type Permissibility for each PS and ESG
    Transaction Code:-OH00
    Personnel Management Personnel AdministrationPayroll DataRecurring Payment and DeductionWage TypesDefine Wage Type Permissibility for each PS and ESG
    Transaction Code:-OH00
    Personnel Management Personnel AdministrationaPPayroll DataAdditional Payments Define Reasons for Changes
    Transaction Code:-OH11
    Personnel Management Personnel AdministrationaPPayroll DataAdditional Payments Wage TypesCreate Wage Type Catalog
    Transaction Code:-OH11
    Personnel Management Personnel AdministrationPayroll DataAdditional Payments Wage TypesCreate Wage Type Catalog
    Transaction Code:-OH00
    Personnel Management Personnel AdministrationPayroll DataAdditional Payments and DeductionWage TypesCheck Wage Type Group Additional Payments
    Transaction Code:-OH13
    Personnel Management Personnel AdministrationPayroll DataAdditional Payments and DeductionWage TypesCheck Wage Type CatalogCheck Wage Type
    Transaction Code:-OH13
    Personnel Management Personnel AdministrationPayroll DataAdditional PaymentsWage TypesCheck Wage Type CatalogCheck Entry Permissibility for Additional Payments
    Transaction Code:-OH13
    Personnel Management Personnel AdministrationPayroll DataAdditional PaymentsWage TypesCheck Wage Type CatalogCheck Wage Type Characteristics.
    Transaction Code:-OH13
    Personnel Management Personnel AdministrationPayroll DataAdditional PaymentsWage TypesCheck Wage Type CatalogCheck Wage Type Characteristics.
    Transaction Code:-OH00
    Personnel Management Personnel AdministrationPayroll DataAdditional PaymentsWage TypesDefine Employee Sub Group Grouping for Primary Wage Type.
    Transaction Code:-OH00
    Personnel Management Personnel AdministrationPayroll DataAdditional PaymentsWage TypesDefine Employee Sub Group Grouping for Primary Wage Type.
    Transaction Code:-OH00
    Personnel Management Personnel AdministrationPayroll DataAdditional PaymentsWage TypesDefine Wage Type Permissibility for each PS and ESG
    Transaction Code:-OH00
    Personnel Management Personnel AdministrationPayroll DataAdditional PaymentsWage TypesDefine Wage Type Permissibility for each PS and ESG
    This is not the complete once just for reference purpose i have given it to u
    Edited by: Sikindar on Nov 19, 2008 9:19 AM

  • What is BI ? How we implement & what is the cost to implement ?

    What is BI ? How we implement & what is the cost to implement ?
    Thanks,
    Sumit.

    Hi Sumit,
                        Below is the description according to ur query
    Business Intelligence is a process for increasing the competitive advantage of a business by intelligent use of available data in decision making. This process is pictured below.
    The five key stages of Business Intelligence:
    1.     Data Sourcing
    2.     Data Analysis
    3.     Situation Awareness
    4.     Risk Assessment
    5.     Decision Support
    Data sourcing
    Business Intelligence is about extracting information from multiple sources of data. The data might be: text documents - e.g. memos or reports or email messages; photographs and images; sounds; formatted tables; web pages and URL lists. The key to data sourcing is to obtain the information in electronic form. So typical sources of data might include: scanners; digital cameras; database queries; web searches; computer file access; etcetera.
    Data analysis
    Business Intelligence is about synthesizing useful knowledge from collections of data. It is about estimating current trends, integrating and summarising disparate information, validating models of understanding, and predicting missing information or future trends. This process of data analysis is also called data mining or knowledge discovery. Typical analysis tools might use:-
    u2022     probability theory - e.g. classification, clustering and Bayesian networks; 
    u2022     statistical methods - e.g. regression; 
    u2022     operations research - e.g. queuing and scheduling; 
    u2022     artificial intelligence - e.g. neural networks and fuzzy logic.
    Situation awareness
    Business Intelligence is about filtering out irrelevant information, and setting the remaining information in the context of the business and its environment. The user needs the key items of information relevant to his or her needs, and summaries that are syntheses of all the relevant data (market forces, government policy etc.).  Situation awareness is the grasp of  the context in which to understand and make decisions.  Algorithms for situation assessment provide such syntheses automatically.
    Risk assessment
    Business Intelligence is about discovering what plausible actions might be taken, or decisions made, at different times. It is about helping you weigh up the current and future risk, cost or benefit of taking one action over another, or making one decision versus another. It is about inferring and summarising your best options or choices.
    Decision support
    Business Intelligence is about using information wisely.  It aims to provide warning you of important events, such as takeovers, market changes, and poor staff performance, so that you can take preventative steps. It seeks to help you analyse and make better business decisions, to improve sales or customer satisfaction or staff morale. It presents the information you need, when you need it.
    This section describes how we are using extraction, transformation and loading (ETL) processes and a data warehouse architecture to build our enterprise-wide data warehouse in incremental project steps. Before an enterprise-wide data warehouse could be delivered, an integrated architecture and a companion implementation methodology needed to be adopted. A productive and flexible tool set was also required to support ETL processes and the data warehouse architecture in a production service environment. The resulting data warehouse architecture has the following four principal components:
    u2022 Data Sources
    u2022 Data Warehouses
    u2022 Data Marts
    u2022 Publication Services
    ETL processing occurs between data sources and the data warehouse, between the data warehouse and data marts and may also be used within the data warehouse and data marts.
    Data Sources
    The university has a multitude of data sources residing in different Data Base Management System (DBMS) tables and non-DBMS data sets. To ensure that all relevant data source candidates were identified, a physical inventory and logical inventory was conducted. The compilation of these inventories ensures that we have an enterprise-wide view of the university data resource.
    The physical inventory was comprised of a review of DBMS cataloged tables as well as data sets used by business processes. These data sets had been identified through developing the enterprise-wide information needs model.
    3
    SUGI 30 Focus Session
    The logical inventory was constructed from u201Cbrain-stormingu201D sessions which focused on common key business terms which must be referenced when articulating the institutionu2019s vision and mission (strategic direction, goals, strategies, objectives and activities). Once the primary terms were identified, they were organized into directories such as u201CProjectu201D, u201CLocationu201D, u201CAcademic Entityu201D, u201CUniversity Personu201D, u201CBudget Envelopeu201D etc. Relationships were identified by recognizing u201Cnatural linkagesu201D within and among directories, and the u201Cdrill-downsu201D and u201Croll-upsu201D that were required to support u201Creport byu201D and u201Creport onu201D information hierarchies. This exercise allowed the directories to be sub-divided into hierarchies of business terms which were useful for presentation and validation purposes.
    We called this important deliverable the u201CConceptual Data Modelu201D (CDM) and it was used as the consolidated conceptual (paper) view of all of the Universityu2019s diverse data sources. The CDM was then subjected to a university-wide consultative process to solicit feedback and communicate to the university community that this model would be adopted by the Business Intelligence (BI) project as a governance model in managing the incremental development of its enterprise-wide data warehousing project.
    Data Warehouse
    This component of our data warehouse architecture (DWA) is used to supply quality data to the many different data marts in a flexible, consistent and cohesive manner. It is a u2018landing zoneu2019 for inbound data sources and an organizational and re-structuring area for implementing data, information and statistical modeling. This is where business rules which measure and enforce data quality standards for data collection in the source systems are tested and evaluated against appropriate data quality business rules/standards which are required to perform the data, information and statistical modeling described previously.
    Inbound data that does not meet data warehouse data quality business rules is not loaded into the data warehouse (for example, if a hierarchy is incomplete). While it is desirable for rejected and corrected records to occur in the operational system, if this is not possible then start dates for when the data can begin to be collected into the data warehouse may need to be adjusted in order to accommodate necessary source systems data entry u201Cre-worku201D. Existing systems and procedures may need modification in order to permanently accommodate required data warehouse data quality measures. Severe situations may occur in which new data entry collection transactions or entire systems will need to be either built or acquired.
    We have found that a powerful and flexible extraction, transformation and loading (ETL) process is to use Structured Query Language (SQL) views on host database management systems (DBMS) in conjunction with a good ETL tool such as SAS® ETL Studio. This tool enables you to perform the following tasks:
    u2022 The extraction of data from operational data stores
    u2022 The transformation of this data
    u2022 The loading of the extracted data into your data warehouse or data mart
    When the data source is a u201Cnon-DBMSu201D data set it may be advantageous to pre-convert this into a SAS® data set to standardize data warehouse metadata definitions. Then it may be captured by SAS® ETL Studio and included in the data warehouse along with any DBMS source tables using consistent metadata terms. SAS® data sets, non-SAS® data sets, and any DBMS table will provide the SAS® ETL tool with all of the necessary metadata required to facilitate productive extraction, transformation and loading (ETL) work.
    Having the ability to utilize standard structured query language (SQL) views on host DBMS systems and within SAS® is a great advantage for ETL processing. The views can serve as data quality filters without having to write any procedural code. The option exists to u201Cmaterializeu201D these views on the host systems or leave them u201Cun-materializedu201D on the hosts and u201Cmaterializeu201D them on the target data structure defined in the SAS® ETL process. These choices may be applied differentially depending upon whether you are working with u201Ccurrent onlyu201D or u201Ctime seriesu201D data. Different deployment configurations may be chosen based upon performance issues or cost considerations. The flexibility of choosing different deployment options based upon these factors is a considerable advantage.
    4
    SUGI 30 Focus Session
    Data Marts
    This component of the data warehouse architecture may manifest as the following:
    u2022 Customer u201Cvisibleu201D relational tables
    u2022 OLAP cubes
    u2022 Pre-determined parameterized and non-parameterized reports
    u2022 Ad-hoc reports
    u2022 Spreadsheet applications with pre-populated work sheets and pivot tables
    u2022 Data visualization graphics
    u2022 Dashboard/scorecards for performance indicator applications
    Typically a business intelligence (BI) project may be scoped to deliver an agreed upon set of data marts in a project. Once these have been well specified, the conceptual data model (CDM) is used to determine what parts need to be built or used as a reference to conform the inbound data from any new project. After the detailed data mart specifications (DDMS) have been verified and the conceptual data model (CDM) components determined, a source and target logical data model (LDM) can be designed to integrate the detailed data mart specification (DDMS) and conceptual data model (CMD). An extraction, transformation and loading (ETL) process can then be set up and scheduled to populate the logical data models (LDM) from the required data sources and assist with any time series and data audit change control requirements.
    Over time as more and more data marts and logical data models (LDMu2019s) are built the conceptual data model (CDM) becomes more complete. One very important advantage to this implementation methodology is that the order of the data marts and logical data models can be entirely driven by project priority, project budget allocation and time-to-completion constraints/requirements. This data warehouse architecture implementation methodology does not need to dictate project priorities or project scope as long as the conceptual data model (CDM) exercise has been successfully completed before the first project request is initiated.
    McMasteru2019s Data Warehouse design
    DevelopmentTestProductionWarehouseWarehouseWarehouseOtherDB2 OperationalOracle OperationalETLETLETLETLETLETLETLETLETLDataMartsETLETLETLDataMartsDataMartsDB2/Oracle BIToolBIToolBIToolNoNoUserUserAccessAccessUserUserAccessAccess(SAS (SAS Data sets)Data sets)Staging Area 5
    SUGI 30 Focus Session
    Publication Services
    This is the visible presentation environment that business intelligence (BI) customers will use to interact with the published data mart deliverables. The SAS® Information Delivery Portal will be utilized as a web delivery channel to deliver a u201Cone-stop information shoppingu201D solution. This software solution provides an interface to access enterprise data, applications and information. It is built on top of the SAS Business Intelligence Architecture, provides a single point of entry and provides a Portal API for application development. All of our canned reports generated through SAS® Enterprise Guide, along with a web-based query and reporting tool (SAS® Web Report Studio) will be accessed through this publication channel.
    Using the portalu2019s personalization features we have customized it for a McMaster u201Clook and feelu201D. Information is organized using pages and portlets and our stakeholders will have access to public pages along with private portlets based on role authorization rules. Stakeholders will also be able to access SAS® data sets from within Microsoft Word and Microsoft Excel using the SAS® Add-In for Microsoft Office. This tool will enable our stakeholders to execute stored processes (a SAS® program which is hosted on a server) and embed the results in their documents and spreadsheets. Within Excel, the SAS® Add-In can:
    u2022 Access and view SAS® data sources
    u2022 Access and view any other data source that is available from a SAS® server
    u2022 Analyze SAS® or Excel data using analytic tasks
    The SAS® Add-In for Microsoft Office will not be accessed through the SAS® Information Delivery Portal as this is a client component which will be installed on individual personal computers by members of our Client Services group. Future stages of the project will include interactive reports (drill-down through OLAP cubes) as well as balanced scorecards to measure performance indicators (through SAS® Strategic Performance Management software). This, along with event notification messages, will all be delivered through the SAS® Information Delivery Portal.
    Publication is also channeled according to audience with appropriate security and privacy rules.
    SECURITY u2013 AUTHENTICATION AND AUTHORIZATION
    The business value derived from using the SAS® Value Chain Analytics includes an authoritative and secure environment for data management and reporting. A data warehouse may be categorized as a u201Ccollection of integrated databases designed to support managerial decision making and problem solving functionsu201D and u201Ccontains both highly detailed and summarized historical data relating to various categories, subjects, or areasu201D. Implementation of the research funding data mart at McMaster has meant that our stakeholders now have electronic access to data which previously was not widely disseminated. Stakeholders are now able to gain timely access to this data in the form that best matches their current information needs. Security requirements are being addressed taking into consideration the following:
    u2022 Data identification
    u2022 Data classification
    u2022 Value of the data
    u2022 Identifying any data security vulnerabilities
    u2022 Identifying data protection measures and associated costs
    u2022 Selection of cost-effective security measures
    u2022 Evaluation of effectiveness of security measures
    At McMaster access to data involves both authentication and authorization. Authentication may be defined as the process of verifying the identity of a person or process within the guidelines of a specific
    6
    SUGI 30 Focus Session
    security policy (who you are). Authorization is the process of determining which permissions the user has for which resources (permissions). Authentication is also a prerequisite for authorization. At McMaster business intelligence (BI) services that are not public require a sign on with a single university-wide login identifier which is currently authenticated using the Microsoft Active Directory. After a successful authentication the SAS® university login identifier can be used by the SAS® Meta data server. No passwords are ever stored in SAS®. Future plans at the university call for this authentication to be done using Kerberos.
    At McMaster aggregate information will be open to all. Granular security is being implemented as required through a combination of SAS® Information Maps and stored processes. SAS® Information Maps consist of metadata that describe a data warehouse in business terms. Through using SAS® Information Map Studio which is an application used to create, edit and manage SAS® Information Maps, we will determine what data our stakeholders will be accessing through either SAS® Web Report Studio (ability to create reports) or SAS® Information Delivery Portal (ability to view only). Previously access to data residing in DB-2 tables was granted by creating views using structured query language (SQL). Information maps are much more powerful as they capture metadata about allowable usage and query generation rules. They also describe what can be done, are database independent and can cross databases and they hide the physical structure of the data from the business user. Since query code is generated in the background, the business user does not need to know structured query language (SQL). As well as using Information Maps, we will also be using SAS® stored processes to implement role based granular security.
    At the university some business intelligence (BI) services are targeted for particular roles such as researchers. The primary investigator role of a research project needs access to current and past research funding data at both the summary and detail levels for their research project. A SAS® stored process (a SAS® program which is hosted on a server) is used to determine the employee number of the login by checking a common university directory and then filtering the research data mart to selectively provide only the data that is relevant for the researcher who has signed onto the decision support portal.
    Other business intelligence (BI) services are targeted for particular roles such as Vice-Presidents, Deans, Chairs, Directors, Managers and their Staff. SAS® stored processes are used as described above with the exception that they filter data on the basis of positions and organizational affiliations. When individuals change jobs or new appointments occur the authorized business intelligence (BI) data will always be correctly presented.
    As the SAS® stored process can be executed from many environments (for example, SAS® Web Report Studio, SAS® Add-In for Microsoft Office, SAS® Enterprise Guide) authorization rules are consistently applied across all environments on a timely basis. There is also potential in the future to automatically customize web portals and event notifications based upon the particular role of the person who has signed onto the SAS® Information Delivery Portal.
    ARCHITECTURE (PRODUCTION ENVIRONMENT)
    We are currently in the planning stages for building a scalable, sustainable infrastructure which will support a scaled deployment of the SAS® Value Chain Analytics. We are considering implementing the following three-tier platform which will allow us to scale horizontally in the future:
    Our development environment consists of a server with 2 x Intel Xeon 2.8GHz Processors, 2GB of RAM and is running Windows 2000 u2013 Service Pack 4.
    We are considering the following for the scaled roll-out of our production environment.
    A. Hardware
    1. Server 1 - SAS® Data Server
    - 4 way 64 bit 1.5Ghz Itanium2 server
    7
    SUGI 30 Focus Session
    - 16 Gb RAM
    - 2 73 Gb Drives (RAID 1) for the OS
    - 1 10/100/1Gb Cu Ethernet card
    - 1 Windows 2003 Enterprise Edition for Itanium
    2 Mid-Tier (Web) Server
    - 2 way 32 bit 3Ghz Xeon Server
    - 4 Gb RAM
    - 1 10/100/1Gb Cu Ethernet card
    - 1 Windows 2003 Enterprise Edition for x86
    3. SAN Drive Array (modular and can grow with the warehouse)
    - 6 u2013 72GB Drives (RAID 5) total 360GB for SAS® and Data
    B. Software
    1. Server 1 - SAS® Data Server
    - SAS® 9.1.3
    - SAS® Metadata Server
    - SAS® WorkSpace Server
    - SAS® Stored Process Server
    - Platform JobScheduler
    2. Mid -Tier Server
    - SAS® Web Report Studio
    - SAS® Information Delivery Portal
    - BEA Web Logic for future SAS® SPM Platform
    - Xythos Web File System (WFS)
    3. Client u2013Tier Server
    - SAS® Enterprise Guide
    - SAS® Add-In for Microsoft Office
    REPORTING
    We have created a number of parameterized stored processes using SAS® Enterprise Guide, which our stakeholders will access as both static (HTML as well as PDF documents) and interactive reports (drill-down) through SAS® Web Report Studio and the SAS® Add-In for Microsoft Office. All canned reports along with SAS® Web Report Studio will be accessed through the SAS® Information Delivery Portal.
    NEXT STEPS
    Next steps of the project include development of a financial data mart along with appropriate data quality standards, monthly frozen snapshots and implementation of university-wide financial reporting standards. This will facilitate electronic access to integrated financial information necessary for the development and maintenance of an integrated, multi-year financial planning framework. Canned reports to include monthly web-based financial statements, with drill-down capability along with budget templates automatically populated with data values and saved in different workbooks for different subgroups (for example by Department). The later will be accomplished using Microsoft Direct Data Exchange (DDE).
    8
    SUGI 30 Focus Session
    As well, we will begin the implementation of SAS® Strategic Performance Management Software to support the performance measurement and monitoring initiative that is a fundamental component of McMasteru2019s strategic plan. This tool will assist in critically assessing and identifying meaningful and statistically relevant measures and indicators. This software can perform causal analyses among various measures within and across areas providing useful information on inter-relationships between factors and measures. As well as demonstrating how decisions in one area affect other areas, these cause-and-effect analyses can reveal both good performance drivers and also possible detractors and enable u2018evidenced-basedu2019 decision-making. Finally, the tool provides a balanced scorecard reporting format, designed to identify statistically significant trends and results that can be tailored to the specific goals, objectives and measures of the various operational areas of the University.
    LESSONS LEARNED
    Lessons learned include the importance of taking a consultative approach not only in assessing information needs, but also in building data hierarchies, understanding subject matter, and in prioritizing tasks to best support decision making and inform senior management. We found that a combination of training and mentoring (knowledge transfer) helped us accelerate learning the new tools. It was very important to ensure that time and resources were committed to complete the necessary planning and data quality initiatives prior to initiating the first project. When developing a project plan, it is important to

  • What is the best way to implement a cluster-wide object ID generator?

    What is the best way to implement a cluster-wide object ID generator?

    What is the best way to implement a cluster-wide
    object ID generator?Always use 3 because it is prime.
    Alternatively more information about the system and the needs of the system might prompt alternative ideas some of which are likely to be better than others for your particular implementation and system constraints.

  • Which Supporting Components do I need if I only want to Install Photoshop from CS4

    I want to install Photoshop only from my Creative Suite (CS4). Which Supporting Components do I need to install?  I only want to install what I need and nothing more.

    When I want to do a custom install, it gives me the option to install these Supporting Components:
    Adobe After Effects CS4 Presets
    adobe CMaps CS4
    Adobe CSI CS4
    Adobe Color - Photoshop Specific CS4
    Adobe Color EU Extra Settings CS4
    Adobe Color JA Extra Settings CS4
    Adobe Color NA Recommended Settings CS4
    Adobe Color Video Profiles AE CS4
    Adobe Color Video Profiles CS CS4
    Adobe Default Language CS4
    Adobe Dynamiclink Support
    Adobe ExtendScript Toolkit CS4\
    Adobe Fonts All
    Adobe Linguistics CS4
    Adobe MotionPicture Color Files CS4
    Adobe Output Module
    Adobe PDF Library Files CS4
    Adobe Photoshop CS4 Support
    Adobe Search for Help
    Adobe Service Manager Extension
    Adobe Type Support CS4
    Adobe Update Manager CS4
    Adobe Utilities CS4
    Adobe WinSoft Linguistics Plugin
    Adobe XMP Panels CS4
    AdobeColorCommonSetCMYK
    AdobeColorCommonSetRGB
    Connect
    PDF Settings CS4
    Photoshop Camera Raw
    kuler

  • What codec do i need to play .mov on windows7

    what codec do i need to play .mov on windows7 i get a message saying im missing components, then sends me to a page with several

    Quicktime should be able to natively handle .mov files. Make sure you have the latest version of Quicktime installed on your computer. If that doesn't work, try an alternative media player like VLC, which can play just about any type of media file.

  • What are the benefits of bw implementation

    Hai Experts,
    If any one have  an idea the below query Please share with me.
      What are the benefits of BW Implementation in an organization.
    Thanks & Regards.
    suresh

    Hi,
    BW/BI is the information backbone engine for SAP and the mySAP landscape. Traditional R/3 reporting does a good job of generating tactical (operational reports). But in the past few years SAP has growingly introduced the business information warehouse as the defacto reporting framework for the mySAP suite of solutions. Some reasons for introducing SAP BW:
    a) An information warehouse that enables analytical reporting
    b) In a typical SAP user community more than 70% of the users would be "read-only" or reporting users. It is unnecessary to provide SAP transaction system reporting to the 70%.
    c) Ability to create reports on the fly providing much more flexibility to information insight.
    BW is clearly the fastest growing SAP technology in terms of user adoption. In the last few years alone, over 3000 installations of BW have been reported. SAP has thrown in a very creative licensing strategy that provides attractive incentives to migrate to mySAP BW.
    The key components of BW that have gained momentum:
    1) Strategic & tactical information analytics
    2) Portal integration for extended and easier information consumption
    3) Pre-delivered best business practices (Business content) for expedited implementations.
    4) BW-BPS/BI IP  for planning and feedback into SAP transaction system
    *pls search forum, you will get  more details**
    Hope this helps,
    Regards
    CSM Reddy

  • What MIBs do I need?

    I'm implementing some testing solution.
    I need to comunicate via SNMP with CISCO MDS 9148 FC switch
    I need to:
    Read interfaces status
    Set interface down
    Set interface up
    The question is what MIBs do I need to be able to do mentioned over? I'm using Python + NetSnmp module.
    Thanks in advance.

    Hi Sergei,
    IF-MIB probably contains what you are looking for to manage interfaces but you will need to load some other mibs as well.
    http://tools.cisco.com/Support/SNMP/do/BrowseOID.do?objectInput=1.3.6.1.2.1.2.2.1.7&translate=Translate&submitValue=SUBMIT
    MIB Loading Order
    http://www.cisco.com/en/US/docs/switches/datacenter/mds9000/sw/5_2/technical/reference/mibs_qkref/MIBRef.html#wp84290

  • What are the pre requisites for implementing ChaRM

    Hi
    We are planning to implement ChaRM & CTS+ in our organization. We are using three system landscape and we have EP, PI, BI and ECC systems.
    Can you please tell me what are the pre requisites for implementing ChaRM & CTS+ and what are the areas i need to look before starting the configuration
    Thanks & Regards
    Anand

    The listed pre-requisites for the implementation of ChaRM are;
    1.  Already completed the basic configuraiton of SAP Solution Manager
    2.  You have set-up your solution, landscape and installed base (iBase) for the systems to be integrated
    3.  you have setup business partners for the users (or at least 1)
    4.  satellite systems are at the minimum  support package level required (note 907768)
    5.  connections are available to the satelite systems
    6.  CTS+ if you wish to control non ABAP change requests through ChaRM

Maybe you are looking for

  • Windows 7 partition won't boot, hangs on black screen / blinking cursor

    Hi all, I just installed windows 7 home premium x64 on a new 100gb partition I made first with bootcamp assistant and then simply with disk utility the second time. I can't get windows to load however, it just sits forever at the black screen with bl

  • Bouncing text

    I have the problem in Adobe Premiere elements 11 with bouncing text (efects "focus" and "blur"). You can see it on video: And if I used these effects, the program is unstable and falling (in editing the effects or in rendering process). Is it problem

  • Zero quantity & but have value

    Dear Gurus Is there any standard report to get the zero quantity but has value. regards sam

  • Need to disable download in firefox, so i can use my download manager

    i want to save download links to my download manager, but firefox downloader is the only option i get

  • AcroRd32.exe process on Application Server Hanging Up

    I run a print job in a reporting system that utilizes Adobe Reader 9.5 for Windows Server 2008. When the job runs, the AcroRd32.exe process hangs up, which prevents me from running anymore print jobs. I cannot end the process via Windows Task Manager