What is the best practice to replicate the change of an object name in HANA AT/AN view where there are dependent views in it?

Dear All HANA Studio Experts,
I'm using SAP HANA Studio Version: 1.0.68 in our current landscape and looking for a expert and intelligent solution for a problem that we have encountered recently.
We have created multiple AT views, AN views and created CA views on top of these AN views to cater different LOB related to our organization. We have also designed BO universe on top of the CA views from HANA and finally created webi reports out of these universes.
In a particular scenario, we have one AT view which is been used by 10 different AN views.
Now we had 3 Requirements like below:
1. To add a new field in this AT view,
2. To update an existing field in this AT view and
3. To remove a field from this AT view which is not required anymore.
Problems Encountered:
Req 1. When we added a new field (e.g. XX_ADD) in the AT view and activate it, its throwing warning that there are dependent objects on this AT view. However no error thrown and it leads to re-work to check if this new field is available in all those dependent views and re-deployed those all.
Question: How to ensure if this addition of object will not impact the rest of the dependent views?
Req 2. When we updated an existing field (e.g. XX_UPDATE_NEW) in the AT view and activate it, its throwing warning that there are dependent objects on this AT view. However the BO reports built on top of these different dependent AN views became inoperative.
Question: How to ensure if we need to update the name of an existing object it will not impact the rest of the dependent AN views?
Req 3. Finally we have removed one object from the existing fields as it is not required. Again we faced warnings that there are dependent AN views on this AT view.
Question: How to ensure if we need to delete an existing object, it will automatically reflect in all dependent AN views?
Please suggest, if it is a good practice to create separate AT,AN and CA views for each of the LOB's and avoiding any dependent views (hence gradually the total no. of logical objects will be increase) or to reuse the AT/AN views whenever possible and in case of any updation/deletion we need to do a huge rework to check and redeploy the dependent objects again and again?
Thanks for reading this detailed explanation and please let me know in case you need any other inputs....don't hesitate to share your expert thoughts on this.

Hi Avishek,
Are you using Development or Modeller Prespective?
Avishek Chakraborty wrote:
Now we had 3 Requirements like below:
1. To add a new field in this AT view,
2. To update an existing field in this AT view and
3. To remove a field from this AT view which is not required anymore.
1) Adding a new Field in the AT View and after activating the view. Then you have to Refresh/Check out the dependent views to see if it is reflecting.
2)  We cannot update the existing field in the AT view i.e like name change if it is being already used in joins in any of the dependent views. It will not allow you to activate
3) To remove the field which is not being used any where, we can make the changes in AT View and Refresh/Check out to see if the dependent views are reflecting the change
Regards,
Krishna Tangudu

Similar Messages

  • What is the best practice for using the Calendar control with the Dispatcher?

    It seems as if the Dispatcher is restricting access to the Query Builder (/bin/querybuilder.json) as a best practice regarding security.  However, the Calendar relies on this endpoint to build the events for the calendar.  On Author / Publish this works fine but once we place the Dispatcher in front, the Calendar no longer works.  We've noticed the same behavior on the Geometrixx site.
    What is the best practice for using the Calendar control with Dispatcher?
    Thanks in advance.
    Scott

    Not sure what exactly you are asking but Muse handles the different orientations nicely without having to do anything.
    Example: http://www.cariboowoodshop.com/wood-shop.html

  • What's the best practice to manage the page file?

           
    We have one Hyper-v Server running windows 2012 R2 with 128 GB RAM and 2 drives (C and D). It setup Automatically manage page file size for all drives. What's the best practice to manage the page file?
    Bob Lin, MCSE & CNE Networking, Internet, Routing, VPN Networking, Internet, Routing, VPN Troubleshooting on http://www.ChicagoTech.net How to Install and Configure Windows, VMware, Virtualization and Cisco on http://www.HowToNetworking.com

    For Hyper-V systems, my general recommendation is to set the page file to 1-4 GB. This allows for a mini-dump should something happen. 99.99% of the time, Microsoft will be able to figure out the cause of the problem from the mini-dump. It does not make
    sense on a Hyper-V system to set aside enough space to capture all the memory on the system because only a very small portion of that memory is used by the parent partition. Most of the memory is under control of the individual VMs.
    Yes, I had one of the Hyper-V product group tell me that I should let Windows manage it.  A couple of times I saw space on my system disk disappear because the algorithm decided it wanted all the space for the page file.  Made it so I couldn't
    patch my systems.  Went back in and set the page file to 1-4 GB and have not had any issues since.
    . : | : . : | : . tim

  • Best practice for replicate the IDM (OIM 11.1.1.5.0) environment

    Dear,
    I need to replicate the IDM production to build the IDM test environment. I have the OIM 11.1.1.5.0 in production with lots of custom codes.
    I have two approaches:
    approach1:
    Manually deploy the code through Jdeveloper and export and import all the artifacts. Issue is this will require a lot of time and resolving dependencies.
    approach2:
    Take the OIM, MDS and SOA schemas export and Import the same into new IDM Test DB environment.
    could you please suggest me what is the best practice and if you have some pointers to achieve the same.
    Appreciate your help.
    Thanks,

    Follow your build document for the same steps you used to build production.
    You should know where all your code is. You can use the deployment manager to export your configurations. Export customized files from MDS. Just follow the process again, and you will have a clean instance not containing production data.
    It only takes a lot of time if your client is lacking documentation or if you re not familiar with all the parts of the environment. What's 2-3 hours compared to all the issues you will run into if you copy databases or import/export schemas?
    -Kevin

  • What are the best practices to extend the overall lifespan of my MacBook Pro and its battery?

    In general what are the recomended practices to extend the lifespan of my batter and other general practice to extend the lifespan and characteristics(such as performance and speed) like new on my MacBook Pro which this past fall (2011)?

    About Batteries in Modern Apple Laptops
    Apple - Batteries - Notebooks
    Extending the Life of Your Laptop Battery
    Apple - Batteries
    Determining Battery Cycle Count
    Calibrating your computer's battery for best performance
    MacBook and MacBook Pro- Mac reduces processor speed when battery is removed while operating from an A-C adaptor
    Battery University
    Kappy's Personal Suggestions for OS X Maintenance
    For disk repairs use Disk Utility.  For situations DU cannot handle the best third-party utilities are: Disk Warrior;  DW only fixes problems with the disk directory, but most disk problems are caused by directory corruption; Disk Warrior 4.x is now Intel Mac compatible. Drive Genius provides additional tools not found in Disk Warrior.  Versions 1.5.1 and later are Intel Mac compatible.
    OS X performs certain maintenance functions that are scheduled to occur on a daily, weekly, or monthly period. The maintenance scripts run in the early AM only if the computer is turned on 24/7 (no sleep.) If this isn't the case, then an excellent solution is to download and install a shareware utility such as Macaroni, JAW PseudoAnacron, or Anacron that will automate the maintenance activity regardless of whether the computer is turned off or asleep.  Dependence upon third-party utilities to run the periodic maintenance scripts was significantly reduced since Tiger.  These utilities have limited or no functionality with Snow Leopard or Lion and should not be installed.
    OS X automatically defragments files less than 20 MBs in size, so unless you have a disk full of very large files there's little need for defragmenting the hard drive. As for virus protection there are few if any such animals affecting OS X. You can protect the computer easily using the freeware Open Source virus protection software ClamXAV. Personally I would avoid most commercial anti-virus software because of their potential for causing problems. For more about malware see Macintosh Virus Guide.
    I would also recommend downloading a utility such as TinkerTool System, OnyX 2.4.3, or Cocktail 5.1.1 that you can use for periodic maintenance such as removing old log files and archives, clearing caches, etc.
    For emergency repairs install the freeware utility Applejack.  If you cannot start up in OS X, you may be able to start in single-user mode from which you can run Applejack to do a whole set of repair and maintenance routines from the command line.  Note that AppleJack 1.5 is required for Leopard. AppleJack 1.6 is compatible with Snow Leopard. There is no confirmation that this version also works with Lion.
    When you install any new system software or updates be sure to repair the hard drive and permissions beforehand. I also recommend booting into safe mode before doing system software updates.
    Get an external Firewire drive at least equal in size to the internal hard drive and make (and maintain) a bootable clone/backup. You can make a bootable clone using the Restore option of Disk Utility. You can also make and maintain clones with good backup software. My personal recommendations are (order is not significant):
    Carbon Copy Cloner
    Data Backup
    Deja Vu
    SuperDuper!
    SyncTwoFolders
    Synk Pro
    Synk Standard
    Tri-Backup
    Visit The XLab FAQs and read the FAQs on maintenance, optimization, virus protection, and backup and restore.
    Additional suggestions will be found in Mac Maintenance Quick Assist.
    Referenced software can be found at CNet Downloads or MacUpdate.
    Be sure you have an adequate amount of RAM installed for the number of applications you run concurrently. Be sure you leave a minimum of 10% of the hard drive's capacity as free space.

  • What's the best way to replicate the contents of servers?

    Here's the situation. Got a client who wants to take his 5 existing servers - one a general file/app server, one an email (probably Exchange) server, one running MS SQL, and I'm not sure about the roles of servers 4 and 5. Needless to say, he wants to take
    the existing servers and relocate them to his home. He also wants to replace the servers with new versions.  The old servers are running an older version of Windows Server (assuming 2003 but could be newer) - but he IS willing to spend money upgrading
    the OS if need be.
    He wants to take the 5 servers and replicate the contents from the old servers to the new ones. Then periodically update the old group with the updated data on the new group - so in case of disaster, he can quickly either move the old servers back to his
    office OR at least recover the data quickly by way of a sync session.
    What is the best way to go about this? Will he need to upgrade the OS on the old servers to make it happen? Money isn't much of an obstacle.

    Here's some info on migrating the file server.
    http://technet.microsoft.com/en-us/library/jj863566.aspx
    For exchange migration I'd ask them here.
    http://social.technet.microsoft.com/Forums/exchange/en-us/home?forum=exchangesvrdeploy&filter=alltypes&sort=lastpostdesc
    The sql should be faily simple backup / restore but better to ask them over here.
    http://social.technet.microsoft.com/Forums/sqlserver/En-US/home?forum=sqlservermigration
    Regards, Dave Patrick ....
    Microsoft Certified Professional
    Microsoft MVP [Windows]
    Disclaimer: This posting is provided "AS IS" with no warranties or guarantees , and confers no rights.

  • What is the best practice to deploy the SharePoint site from test to production environment?

    We are beginning to start a new SharePoint 2010 and 2013 development projects, soon developing new features, lists, workflows, customizations to the SharePoint site, customization to list forms and would like to put good practice (that will help in deployment)
    in place before going ahead with development.
    What is the best way to go about deploying my site from Development to Production?
    I am using Visual Studio 2012 and also have Designer 2013...
    I have already read that this can be done through powershell, also through visual studio and also via designer. But at this point I am confused as to which are best practices specifically for lists, configurations; workflows; site customizations; Visual studio
    development features; customization to list forms etc. You can also provide me reference to links/ebook covering this topic.
    Thanks in advance for any help.

    Hi Nachiket,
    You can follow below approach where the environments has been built in similar fashion
    http://thesharepointfarm.com/sharepoint-test-environments/
    if you have less data then you can use  http://spdeploymentwizard.codeplex.com/
    http://social.technet.microsoft.com/Forums/sharepoint/en-US/b0bdb2ec-4005-441a-a233-7194e4fef7f7/best-way-to-replicate-production-sitecolletion-to-test-environment?forum=sharepointadminprevious
    For custom solutions like workflows etc you can always build the WSP packages and deploy across the environments using powershell scripts.
    Hope this helps.
    My Blog- http://www.sharepoint-journey.com|
    If a post answers your question, please click Mark As Answer on that post and Vote as Helpful
    Hi, can you answer me specifically with regards to the foll:-
    lists
    configurations
    workflows
    site customizations like changes to css/masterpage
    Visual studio webparts
    customization to list forms
    Thanks.

  • What are the best practices for using the enhancement framework?

    Hello enhancement framework experts,
    Recently, my company upgraded to SAP NW 7.1 EhP6.  This presents us with the capability to use the enhancement framework.
    A couple of senior programmers were asked to deliver a guideline for use of the framework.  They published the following statement:
    "SAP does not guarantee the validity of the enhancement points in future releases/versions. As a result, any implemented enhancement points may require significant work during upgrades. So, enhancement points should essentially be used as an alternative to core modifications, which is a rare scenario.".
    I am looking for confirmation or contradiction to the statement  "SAP does not guarantee the validity of enhancement points in future releases/versions..." .  Is this a true statement for both implicit and explicit enhancement points?
    Is the impact of activated explicit and implicit enhancements much greater to an SAP upgrade than BAdi's and user exits?
    Is there any SAP published guidelines/best practices for use of the enhancement framework?
    Thank you,
    Kimberly
    Edited by: Kimberly Carmack on Aug 11, 2011 5:31 PM

    Found an article that answers this question quite well:
    [How to Get the Most From the Enhancement and Switch Framework as a Customer or Partner - Tips from the Experts|http://www.sdn.sap.com/irj/scn/index?rid=/library/uuid/c0f0373e-a915-2e10-6e88-d4de0c725ab3]
    Thank you Thomas Weiss!

  • Where is the best practice to have the controlling logic in scenarios....

    The following steps need to executed in the system:
    1> Need to store the data that User has selected in the screens that he has visited till now (selecting data is not part of this step 1 for now); focus of step 1 is there is some data which needs to be stored in session.
    2> After keeping data in session, the system needs to check and if current user's account is Not created then force the user to create user account;
    3> The system then checks If user is Not logged in. If not forces the user to login;
    4> Process the data that has been searched and selected in step 1;
    EJBs are used for providing the solution;
    Now, the problem i have been having is where should the controlling logic (basically the block which invokes the above 4 steps in sequence) needs to be sitting?
    Should it be in the EJB tier or WEB tier?
    If we decide to store the data in the session bean, Should data get stored in session bean and then should control go back to front controller servlet in web tier;
    Should the enclosing function making checks for steps 2 and 3 and followed by invoking the step 4 be in Front controller servlet or in business delegate or in stateful session bean?
    Any thoughts or any useful links to let me know what is feasible / not feasible.
    I mean, whether such controlling logic which requires ui to be brought up and then control come back can never be sitting in session bean -- is feasible or not.
    Thanks in advance.
    Message was edited by:
    VMR
    Message was edited by:
    VMR

    (Assuming you mean Essex in the UK), go here https://selfsolve.apple.com/GetWarranty.do and enter your iPod's serial number to start with. (Change the country in the drop-down menu if you're not in the UK.) This will guide you to an Authorised Apple Service Provider (AASP) local to you.

  • What are the best practice to design Dashboard?

    Hi Friends,
    What are the best practice to design Dashboard?
    Thanks

    1) Use section name instead of Title View to name your report. This way sections can be collapsed and user can still see report name.
    2) Enable alternate coloring in tables and pivots for easy readablity and set table and pivot widths to 100% (for reports in dashboards) to reduce white space and achieve a more "professional look."
    3) Use column selectors and view selectors to reduce the width of reports and reduce the amount of columns user sees to a "practical minimum."

  • What is the best practice to get thumbnail of photo which I get from cameraUI

    Hi,
    I building a photo sharing application in android and I want to upload a smaller size of photo just after I receive the photo from the cameraUI.
    What is the best practice to resize the photo and to upload it, At the moment I doesn't resize it And the file size is big (1MB) for HTC DESIRE HD ?
    Is it possible to resize the image and keep the EXIF data ?
    Thanks,
    Nimrod.

    Yep,
    Media Manager is the way to go.
    Read the manual aout it.
    It is all explained in there.
    Rienk

  • What is the best practice to create IDM user and target accts via recon

    usecase:
    LDAP<--->idm---->AD.
    User exists in LDAP. IDM and AD are empty. Need to create IDM user and AD acct from LDAP data.
    I can recon against LDAP and create the IDM user. But I cannot create AD acct in the same recon process. What is the best practice to do the above.

    i think you have to have a "Per-account Workflow" set which creates the user in AD.

  • What is the best practice for logging runtime error  or uncheked exceptions

    hello
    my main problem is logging the nullPointerException to a file
    suppose I have this method
    public void myMethod()
       //..... some declaration here
       User user = obj.findUser("userx"); //this may return null
       System.out.println("user name is "+user.getName()); // I may have a null pointer exception here
    }so what is the best practice to catch the exception ??
    can I log the exception without catching it ???
    thank you

    A terrible way of logging exceptions.Not that im not agreeing with you, but why? (I havent
    actually used this)Because it's always on, for one thing. It's not really configurable either, unless you go to some trouble to make it so. You'd have to provide an InputStream. How? Either at compile-time, which is undesirable, or by providing configuration, which a logging framework has already done, better. Then there's the fact that you can't log anything other than the stacktrace - not particularly helpful a lot of the time. In short, it's a buggy and incomplete solution to something that's already been solved much much better by, for example, Log4J

  • What is the best practice to protect coldfusion administrator login page

    Hi all,
    Can someone suggest what is the best practice to protect the administrator login? At the moment, there is only the normal administrator page password to protect. It seems like not very secure especially when the application is on the internet.
    Regards,
    Bubblegum.

    You can protect the page with file system level privs.  Setup a new virtual server that maps to a seperate copy of /cfide (and remove /admin and /adminapi from the other cfide folder your internet sites use).  Limit what IP addresses can hit /cfide.
    We run multiple instances, so we connect directly to each instance to manage it.  And those ports aren't accessable on the internet.  To top it off, we have an ISAPI ReWrite rule that sends a 404 if you try /cfide/administrator or adminapi.
    If you're using CF8, you can set it up so it requires a specific username instead of a generic name.

  • What is the best way to mimic the data from production to other server?

    Hi,
    here we user streams and advanced replication to send the data for 90% of tables from production to another production database server. if one goes down can use another one. is there any other best option rather using the streams and replication? we are having lot of problems with streams these days they keep break and get calls.
    I heard about data guard but dont know what is use of it? please advice the best way to replicate the data.
    Thanks a lot.....

    RAC, Data Guard. The first one is active-active, that is, you have two or more nodes accessing the same database on shared storage and you get both HA and load balancing. The second is active-passive (unless you're on 11.2 with Active Standby or Snapshot Standby), that is one database is primary and the other is standby, which you normally cannot query or modify, but to which you can quickly switch in case primary fails. There's also Logical Standby - it's based on Streams and generally looks like what you seem to be using now (sort of.) But it definitely has issues. You can also take a look at GoldenGate or SharePlex.

Maybe you are looking for