Azure Account Management Best Practices

Hello,
Hoping someone will be able to help me or point me in the right direction. First, a general overview of what I'm trying to do:
I am trying to utilize the IaaS portion of Azure to setup my application on a Windows VM.  Each customer will have their own individual virtual machine configured specifically for them, running my application that they can remote into and use.
My thought was that I would create a separate Subscription in Azure for each customer. That way, I could manage the billing information more easily. Within each Subscription would be all of the Azure resources for that customer: VMs, Storage Accounts, Backup
Vaults, Virtual Networks, etc. Since there is no limit to the number of subscriptions per account, this seems like it would also scale well.
My question is...is this the best way to do this? Since these customers will really have nothing in common besides using my application, I think creating separate Subscriptions with separate Virtual Networks would be the best way to keep them completely
isolated from one another. Does anyone see any problem with this approach or know or a better way to do this? Is there a guide or blog post somewhere that details a similar scenario and how to approach it?
Thanks in advance

Hi,
Yes, you are right.
Best practise would be to create new accounts for your employees usage.
But there are many other ways to do this apart from the one you says.
As the services grow the best way to track is by creating new accounts and deleting the ones when his employees leave the organization.
Please have a look into these articles:https://technet.microsoft.com/en-us/magazine/gg607453.aspx
https://msdn.microsoft.com/en-in/library/azure/hh531793.aspx
Regards,
Azam khan

Similar Messages

  • Solution Manager best practices about environments

    Hello,
    we intend to use Solution Manager 4.0.
    My question : I wonder whether we need to have a single instance a SM (production) or whether we need to have multiple instances (one development SM where developments and customizing will be performed and one production SM populated with transport requests coming from the development SM) ?
    What are the best practices ?
    Thank you.
    Regards,
    Fabrice

    Dear Fabrice,
    In principle you donot need to have 2 instances of Solution Manager. 1 Instance is sufficient enough for monitoring all the Satellite Systems.
    However if you intending to have Customized ABAP on Solution Manager then it might be a good idea to do so in an different client in the same instance keeping the client as an development client.
    Most of the customizing in Solution Manager is non transportable hence it should be directly done in the productive client.
    Hope this answers your queries.
    Regards
    Amit

  • What are project management best practices?

    I created a test project in Premiere Elements 12, and saved it in a directory named "Michaels Posters".   Then I archived the project to this directory and it created a "Copied_My\ new\ video\ project1" directory with all of the media files.  Then I added a video clip to the project, archived it again, and it created the "Copied_My\ new\ video\ project1_001" folder below.
    My first real project will be a video highlights video of my 4 years old for 2013.  This will involve editing the same project several nights a week, for maybe a couple of months.  This would result in numerous "Copied_My\ new\ video\ project1_NNN" directories being created, assuming I archive the project each night.
    So what is the best practices for managing a larger project like this, and avoid using a lot of disk space for the same project?
    Michaels\ Posters/
    ├── Adobe\ Premiere\ Elements\ Preview\ Files
    │   └── My\ new\ video\ project1.PRV
    ├── Copied_My\ new\ video\ project1
    │   ├── Adobe\ Premiere\ Elements\ Preview\ Files
    │   ├── Encoded\ Files
    │   └── Layouts
    ├── Copied_My\ new\ video\ project1_001
    │   └── Adobe\ Premiere\ Elements\ Preview\ Files
    ├── Encoded\ Files
    │   └── My\ new\ video\ project1.prel
    ├── Layouts
    └── Media\ Cache\ Files

    I do work with the LAST archived project file, which contains ALL necessary resources to edit the video.  But then if I add video clips to the project, these newly added clips are NOT in the archived project, so I archive it again.
    The more I think about it, the more I like this workflow.  One disadvantage as you said is duplicate videos and resource files.  But a couple of advantages I like are:
    1. You can revert to a previous version if there are any issues with a newer version, e.g., project corruption.
    2. You can open the archived project ANYWHERE, and all video and resource files are available.
    In terms of a larger project containing dozens of individual clips like my upcoming 2013 video highlights video of my 4  year old, I'll delete older archived projects as I go, and save maybe a couple of previous archived projects, in case I want to revert to these projects.
    If you are familiar with the lack of project management iMovie, then you will know why I am elated to be using Premiere Elements 12, and being able to manage projects at all!
    Thanks again for your help, I'm looking forward to starting my next video project.

  • Multiple room management -- best practice -- server side http api update?

    Hi Folks, 
    Some of the forum postings on multiple room management are over year old now.  I have student/tutor chat application which has been in the wild for 5 months now and appears to be working well.  There is a single tutor per room, multiple chats and soon to be a whiteboard per student, which is shared with the tutor in a tabbed UI. 
    It is now time to fill out the multiple tutor functionality, which I considered and researched when building, but did not come to any conclusions.   I'm leaning towards a server side implementation.  Is there an impending update to the http api?
    Here is what I understand to be the flow:
    1) server side management of who is accessing the room
    2) load balance and manage the room access 1 time user and owner session from the server side
    3) for my implementation, a tutor will need to login to the room, in order for it to be available
    4) Any reconnection would in turn need to be managed by the server side, and is really a special case of room load balancing.
    My fear is that at some point I'm going to need access to the number of students in the room or similar and this is not available, so that I'll need client functionality, which will need update the server side manager.
    As well, I'm concerned about delays on the server side access to which might create race conditions in a re-connect situation.  User attempts to reconnect, but server side manager thinks that the user is already connected.
    Surely this simple room management has been built, does anyone have any wisdom they can impart?  Is there any best practice guidance or any samples?
    Thanks,
    Doug

    Hi Raff, Thanks a ton for the response.
    I wasn't clear on what I was calling load balancing.  What I mean by this is room assignment for student clients.  We have one tutor per room.  There are multiple students per room, but each is in their own one-on-one chat with the tutor.
    I'm very much struggling with where to do the room assignment / room managemnt, on the server side or on the client side (if that is even possible).  In my testing it is taking upwards of 10 seconds minimum to get a list of rooms (4 virtually empty rooms) and to query the users in a single room (also a minimum of users/nodes in the queried room).   If after this point, I 'redirect' the student to the least full room, then the student incurs the cost of creating a new session and logging into the room.  As well I intend to do a bit of xml parsing, and other processing, so that 10 seconds is likely to grow.
    Would I see better performance trying to do this in the client?
    As far as the server side, at what point does a room go to 'not-active'?
    When I'm querying the roomList, I am considered one of the 'OWNER' users in the UserLists.  At what point can it be safe to assume that I have left the room? 
    Is there documentation on the meaning and lifecycle of the different status codes?  not-active,  not-running, and ok?  Are there others?
    How much staleness can I expect from the server-side queries?
    As far as feature set, the only thing that comes to mind is xpath and or wild card support for getNode() but i think this was mentioned in other posts.
    Regarding the reconnection issues, I am timing out the student after inactivity, and this is probably by and large the bulk of my reconnect use cases.  This and any logout interaction from the student presents a use case where I  may want reassign the student return to the same room as before.  I can envision scenarios of a preferred tutor if available etc.  In this case, I'll need to know list of rooms.  In terms of reconnection failover, this is not not a LCCS / FMS issue.
    Thanks again for responding.

  • SRM EBP User management - best practice followed for your customer.

    Hello All,
    What are the best practices followed on SRM User manageemnt  followed for your customers.
    (1)When employee/ buyer  leave the organisation  ? what actions you do ? do you lock the users?
    (2) If any thing interested share your experiences.
    (3) What exactly customer expects from SRM systems on SRM user management?
    (4) SAP audit/ customer audit practice on USER management ?
    Any piece of information on your experiece/ best practice  is appreciated.
    regards
    Muthu

    Thanks Peter .
    it is happening only in SRM right.
    Is any work around for this issue.
    In future SRM any planing to take care of this.
    in ECC i can delete the user whenever the user moves .
    All SRM customers will be very happy if SRM gives some workaround for this issue.
    Every customer wants to reduce cost .
    How can I find what are the opening documents for this user in one shot ?
    thanks for answering this question.
    I have seen our Eden Kelly report helps for shopping cart and other BO.
    You are doing good job on our SRM  WIKI innovative topics and discussons. I appreciate.
    why i am raising this concern is that one user left the organisation and again we want to edit the data whch entered by the left user . system will not allow us to do after deleting the user.
    so we are approaching SAP for ths help.
    It is very difficult to convice the customers on this issues.
    br
    muthu

  • Hotfix Management | Best Practices | WCS | J2EE environment

    Hi All,
    Trying to exploit some best practices around hotfix management in a J2EE environment. After some struggle, we managed to handle the tracking of individual hotfixes using one of our home-grown tools. However, the issue remains on how to manage the 'automated' build of these hotfixes, rather than doing the same manually, as we are currently doing.
    Suppose we need to hotfix a particular jar file on a production environment, I would need to understand how to build 'just' that particular jar. I understand we can label the related code (which in this case could be just a few java files). Suppose this jar contains 10 files out of which 2 files need to be hotfixed. The challenge is to come up with a build script which builds -
    - ONLY this jar
    - the jar with 8 old files and 2 new files.
    - the jar using whatever dependent jars are required.
    - the hotfix build script needs to be generic enough to handle the hotfix build of any jar in the system.
    Pointers, more in line with a WCS environment would be very much appreciated!
    Regards,
    Mrinal Mukherjee

    Moderator Action:
    This post has been moved from the SysAdmin Build & Release Engineering forum
    To the Java EE SDK forum, hopefully for closer topic alignment.
    @)O.P.
    I don't think device driver build/release engineering is what you were intending.
    Additionally, your partial post that was accidentally created as a duplicate to this one
    has been removed before it confuses anyone.

  • Working with version management and promotion management best practices BO 4.1

    Hi Experts
    I wondered if anybody knows if there is a document or something about best practices to work with the version management and promotion management in BO 4.1?
    Our Environment includes two servers. The first one is our development and test server. The second server is our prod system.
    Now on the dev server we have basically two folders called dev and test. We control access to them with a right system based on the folder structure.
    My question now is how you would work in this scenario (third server is not an option). Main target is to have as few reports as possible. Therefore we try to work with the version management system and only have one version of each report in the dev folder of the cms. But this is where problems start. Sometimes the newest version is not the version we want to publish to the test folder or even prod server.
    How would you publish the report to the other folder? Make a copy of the concerned report (transport to the same system via promotion management is not possible). Also how would you use the version management in regards to the folder structure? Only use version management in dev folder and export reports to test folder (out of vms control) or also use vms in test folder and how would that work?
    Further more I’d be interested in learning best practices with promotion management. I found out that the promotion of a report that doesn’t exist in prod doesn’t make any problems. But as soon as an older version already exists there is only partial success and the prod folder gets renamed to “test”.
    Any suggestions on how to handle these problems?
    Thank you and regards
    Lars

    Thank you for your answer.
    So you are basically proposing to work with the vms in the dev folder and publish the desired version to the test folder. And the test folder is out of version control in this scenario if I understood you correctly (like simple data storage)?
    And how would you suggest promoting reports to the prod system? Simply by promoting the
    desired version from dev folder directly to prod? This would probably lead to inconsistence because we would need to promote from dev system to test and dev to prod instead of promoting a straight line from dev over test to prod. Furthermore it would not solve the problem of the promoting result itself (A new folder called dev will be generated in prod but the report gets promoted to the prod folder if there was no report before).
    Thank you for the link. I came across this page just a few days ago and found also lots
    of other tutorials and papers describing the basic promoting process. The promoting process in general is clear to me but I wondered if it is possible to change some parameters to  prevent folder renaming for example.
    Regards
    Lars

  • Azure table design best practice

    What's the best practice for designing tables in Azure Tables to optimize query performance?

    Hi Raj,
    When we design the azure table, we need to consider the scalability of the azure table.
    and selecting the PartitionKey is very more important to scalability.
    Basically, we have two options which have their advantages and disadvantages:
    One Option: having a single partition by having the same value for PartitionKey for all entities to
    Second Option: having a unique value for PartitionKey for every entity
    More information about how to get the most out of windows azure tables ,please refer to the link below:
    http://blogs.msdn.com/b/windowsazurestorage/archive/2010/11/06/how-to-get-most-out-of-windows-azure-tables.aspx
    There is  also a detailed article which explain how to design a scalable partitioning strategy for Windows Azure Storage,please refer to the link below:
    http://msdn.microsoft.com/en-us/library/hh508997.aspx
    Best Regards,
    Kevin Shen

  • X2100M2 Embedded LIghts Out Manager best practice

    Hi guys,
    I'm in worry about the best practice of configuring network interface on a X2100M2 Solaris 5.10 for the Embedded LIghts Out Manager. Hope you can help. I haven't find any documents of it which explain the best practice.
    Here is the situation :
    I've have 4 network interfaces but I only need two of them. So I decide to use the bge0 and the bge1 interfaces.
    bge0 is the server interface with an IP with .157
    bge1 is the ELOM interface with an IP with .156
    In the past, it was the reverse : bge0 was the ELOM with .156 and bge1 was the network server int. with .157
    Could you please guys let me know what is the best practice? Does the int.0 must be the server one? Is it possible to have network problem with this kind of configuration?
    Thanks
    Cheers,

    hi guys,
    No one have a clue? I've got some dukeDolars to offer...
    Tks

  • Operating system image build and management best practices?

    how do we create gold images for servers/desktops,
    Best practices image management,
    How do we control changes?
    How do we prevent unauthorized changes (installation of software)?
    What tools we can use for above.

    I use MDT 2013 Lite Touch to create my images
    http://www.gerryhampsoncm.blogspot.ie/2014/03/create-customised-reference-image-with.html
    You should use in-built ConfigMgr Role Based Access Control to manage images afterwards (look at the Operating System Deployment Manager role).
    Gerry Hampson | Blog:
    www.gerryhampsoncm.blogspot.ie | LinkedIn:
    Gerry Hampson | Twitter:
    @gerryhampson

  • ODP + Client Version Management Best Practices

    I am working with a client who is using the Oracle Developer Tools for Visual Studio to develop their application. The IT folks deploy a new server with the latest version of the Oracle client (i.e. 11.1.0.7.0) and the developers are using the latest ODT (i.e. 11.1.0.7.20). As such the publisher policy never gets a chance to add any value. Because of the unusual versioning scheme we always end up having to copy the Oracle.DataAccess.dll from the Oracle client into the deployment folder and add an assembly binding redirect. If we do not we get the following exception:
    The provider is not compatible with the version of Oracle client
    As you might expect this is VERY annoying to deal with as it adds complexity to our deployment process. We use TNS names to connect and those files are stored in the client\network\admin folder as well. Can someone please help me clear up what the best practice is to deploy an application developed using ODT/ODP.NET so that we don't have to hack the Oracle.DataAccess.dll every time?
    Thanks,
    Colin

    Thank you for your answer.
    So you are basically proposing to work with the vms in the dev folder and publish the desired version to the test folder. And the test folder is out of version control in this scenario if I understood you correctly (like simple data storage)?
    And how would you suggest promoting reports to the prod system? Simply by promoting the
    desired version from dev folder directly to prod? This would probably lead to inconsistence because we would need to promote from dev system to test and dev to prod instead of promoting a straight line from dev over test to prod. Furthermore it would not solve the problem of the promoting result itself (A new folder called dev will be generated in prod but the report gets promoted to the prod folder if there was no report before).
    Thank you for the link. I came across this page just a few days ago and found also lots
    of other tutorials and papers describing the basic promoting process. The promoting process in general is clear to me but I wondered if it is possible to change some parameters to  prevent folder renaming for example.
    Regards
    Lars

  • CM (Configuration Managment) Best Practices for DBs?

    Hi:
    I've got a question on a topic I haven't seen discussed before. I have a project with application code and PL/SQL code and an Oracle 10g database that has triggers, constraints, etc. For a Java or C# project the concept of CM and a central code repository works, largely because the CM tool is able to enforce limits on who can do what to a given resource over time. Developers MUST check out the source files before modifying and adding back to the CM tree.
    Database objects are more on an honor system. Even if you put the code for the tables, triggers, constraints, packages etc. under CM, nothing prevents a developer or multiple developers from modifying those objects in the development database and never updating the create scripts that are in the CM tree. So given that there are thousands of systems out there with Java/C# or whatever front ends and PL/SQL back ends in databases that have all sorts of DB objects in them, what are the best CM practices? How does a developer know the package he's about to modify is the last one that got pushed to the production server?
    Constraints are problematic in several unique ways. They can be created in several places by different syntax (in the table create command; as a stand alone command after the table exist; named or system generated name if defined in the table definition). Things like SQL Developer and Toad can create code trees that contain all the scripts needed to build a database from scratch and a top level build script to run them all, but what's the best structure for that? Do table create scripts JUST define columns and separate scripts create all constraints?
    I like to think someone has crossed this particular bridge before and has some answers to my questions. Any takers?
    Thanks.

    Gaff wrote:
    No. With non-DB code, it isn't voluntary. The developer needs to check a file in or out before modifying it and the CM system enforces either that only one person at a time can edit the file or if multiple ones can, the merging of the results. Once the PL/SQL package is in a database, a developer can change it and NOT change the underlying file. I don't agree that there is a difference. All source control systems I have seen allow a copy to be taken or the source to be viewed without checking it out. I have a checked out source file on my PC. I copy it. I check in my changes, I make changes to the copy. I have made a change without changing the checked in source. Just like I could in the database. Should this change make it into the final build? No, and I can't complain when it doesn't, just like the unauthorized change directly in the database is overwritten and lost. I see no difference.
    Making changes to database objects directly in the database is a terrible practice, you can enforce the same lack of tolerance for it as you would for any other developer circumventing the source control and deployment process in any other language.
    - You tell them not to do it.
    - You overwrite their changes if they do do it.
    - You fire them if they continue to do it.
    >
    I think the solution is that by default DB developers have read-only access to the objects. When they identify the pieces that need to change to fix a given bug, some 3rd party (Software Lead? DBA?) grants them CRUD on the identified objects and verifies that their modified code makes it into the CM tree when the bug is fixed.The developer can make their own changes in the Dev database. Changes to the Test database prior to production deployment can only be made by the DBA using scripts from the source control system.
    This works flawlessly.

  • Software Change Management best practices

    Hi All,
    I am curious how most people here are managing the changes that they make to their SAP code that is running on the IBM i.
    For example, how do you perform version control? Multiple release management? Task management? How do you manage approvals for promotion? How do you comply with SOX "separation of duties" (i.e. someone different has to promote to production).
    Thanks for any insight here!
    Joe.

    Hi All,
    I am curious how most people here are managing the changes that they make to their SAP code that is running on the IBM i.
    For example, how do you perform version control? Multiple release management? Task management? How do you manage approvals for promotion? How do you comply with SOX "separation of duties" (i.e. someone different has to promote to production).
    Thanks for any insight here!
    Joe.

  • Photo file management - best practice query

    A number of recent posts have discussed file management protocols, but I wondered if some of the experts on this forum would be so kind as to opine on a proposed system set up to better manage my photo files.
    I have an imac, time machine & various external hard drives.
    I run Aperture 3 on my imac (my main computer), with about 15k referenced images. Currently my photo masters are kept on my imac, as is my aperture library file, and vault. After editing in APerture, I then export the edited jpegs onto another part of my imac. The imac is backed up to time machine and an off site drive.
    Following some of the threads, the main message seems to be to take as many of my photo files as possible off my imac, and store on a separate drive. So does the folloing set up sound better?
    *Aperture run on imac, still using referenced images
    *Master images moved from imac to external drive 1
    *Aperture library file moved from imac to external drive 1
    *Aperture vault moved to external drive 1
    *External drive 1 backed up to external drive 2. Run idefrag on both drives regularly.
    *Edited exports from Aperture kept on imac, which then feed Apple TV, iphone, mobileme etc. Backed up to time machine.
    *If ever I ran Aperture on an additional machine, presumably I could just plug and play / synch with external hard drive 1.
    Is that a "good" set up? Any enhancements? The set up would seem to free up my boot volume, whilst hopefully maintaining the satefy / integrity of my files. But happy to be told it is all wrong!!
    Many thanks
    Paul

    Seems to be a good approach. However,
    Depending on how much disk space on the local drive and the speed of the external along with the iMac specs... you might keep the the library on the iMac instead of the external, but keep the masters on the external. Assuming referenced, then you could keep the vault on the external (as is I wouldn't put vault on the same external drive as both the masters and library).

  • Nexus 7000 - Management best practice in multi-VDC environment

    Hello all, my topology includes an Admin VDC, a Core\Agg VDC and a Storage VDC for FCoE traffic. In the absence of a dedicated OOB Management switch, could I patch from an interface in the Core\Agg VDC to the Mgmt0 interface of the SUP2 and get access to mgmt0 IP of all VDCs? Or is an OOB management switch a necessity to manage this topology?  
    Any information would be appreciated.
    Kind regards
    Rays

    Thanks for the reply Richard. I understand the risk with not having an OOB network, thanks for that.
    So without the OOB management network, are you saying I need a separate switch to which I can connect a physical interface from each VDC to, for management purposes? And to access the Storage and Admin VDC, would I connect the mgmt0 interfaces to the same switch? As i understand it you cannot allocate a physical i/f to the Admin VDC, so the only method is via console or mgmt0 interface...
    Thanks for you're assistance...
    Rays

Maybe you are looking for

  • Regarding Product group Creation

    Dear SAP Guru                           I am facing issues while creating Product group by the code MC*$.The error is "The field  is defined as a required field; it does not contain an entry Message no. MG144 ".Also while cretaing through the materia

  • Text in Report Painter not displayed in the logon language.

    Dear fiends, I need change the language in the Report Painter, but i don't get translate the texts. My logon logon language is Spanish but in the report painter the language is English. In the transaccition GCRS the laguage is Spanish. I copy these r

  • How do I know if my T400 is with or without IEEE 1394?

    My Lenovo automated diagnostics determined that the BIOS software on my T400 is flawed, and that I should download a replacement BIOS.  The page that I am led to gives me a choice of two different BIOSes:   http://www-307.ibm.com/pc/support/site.wss/

  • Re: Can't increase monthly payments online

    SORRY HAD TO COME THROUGH YOUR POST TO ASK QUESTION DONT REALLY UNDERSTAND HOW TO USE FORUM . ANYWAY 6 JUNE CONTACTED BT ABOUT CHANGING BILL DATE FROM 22 TO 30 OF MONTH SAID I COULDNT DO IT COS I PAY IN FULL EVERY MONTH ANYWAY THEY SUGGESTED I CHANGE

  • Hooking iMac 20" To 42" LCD (Specific LCD Brand)

    Good Morning, I would love you guys to help me today! I just bought a new iMac 20" as well as a Dynex 40" 1080dpi LCD HDTV. I would love to hook up my iMac dispay to the LCD screen. I've scanned through the forum to get a better idea of how to connec