Clarification regarding RMAN usage

Hi refering to http://www.dba-oracle.com/real_application_clusters_rac_grid/recovery_catalog.html
i have two DBs one CATDB(on oracle 9i) and other PRODdb(on 10g)
now i have created recovery catalog in CATDB .
My doubt is while registering the database. from which machine shall i run rman (machine having 10g or 9i or can i invoke RMAn from either of them).
rman TARGET sys/oracle@PRODdb CATALOG rman/cat@catdb
in the above given step i have doubt. which RMAN to use? 9i or 10g or it does not matter.
Thanks

user8896122 wrote:
Hi refering to http://www.dba-oracle.com/real_application_clusters_rac_grid/recovery_catalog.html
i have two DBs one CATDB(on oracle 9i) and other PRODdb(on 10g)
now i have created recovery catalog in CATDB .
My doubt is while registering the database. from which machine shall i run rman (machine having 10g or 9i or can i invoke RMAn from either of them).
rman TARGET sys/oracle@PRODdb CATALOG rman/cat@catdb
in the above given step i have doubt. which RMAN to use? 9i or 10g or it does not matter.
ThanksYou should run this command from the server which you take a backup of it
[Kamran Agayev A. |http://kamranagayev.wordpress.com] (OCP 9i/10g)
Author of the upcoming book - *"Oracle Backup & Recovery: Expert secrets for using RMAN and Data Pump"*
http://www.rampant-books.com/book_1002_rman_backup_recovery.htm

Similar Messages

  • Needs Clarification Regarding Segments and Datafiles

    Hi,
    I want clarification regarding Segments, Datafiles and extents.
    As we know that A segment is made of one or more extents and extents are composed of one or more datablocks in the HD.
    Since all data are store in Datafiles which are composed of extents and datablocks. I want to know weather a table(Segment) can span to multiple datafiles or in a sigle datafile.
    Regards,
    D.Abbasi

    And an easy way to check it by yourself :
    SQL> create tablespace abbasi_tbs
      2  datafile 'E:\ORADATA\DEMO111P\abbasi_01.dbf' size 1m autoextend off,
      3           'E:\ORADATA\DEMO111P\abbasi_02.dbf' size 1m autoextend off;
    Tablespace created.
    SQL> create table abbasi_tbl (id number)
      2  tablespace abbasi_tbs;
    Table created.
    SQL> insert into abbasi_tbl
      2  select rownum as rn
      3  from   dual
      4  connect by level <=10000;
    10000 rows created.
    SQL> commit;
    Commit complete.
    SQL> select distinct file_id
      2  from   dba_extents
      3  where  segment_name ='ABBASI_TBL';
       FILE_ID
             6
             7
    or...
    SQL> select distinct DBMS_ROWID.ROWID_RELATIVE_FNO(rowid)
      2  from   abbasi_tbl;
    DBMS_ROWID.ROWID_RELATIVE_FNO(ROWID)
                                       6
                                       7
    SQL> select file_name from dba_data_files where file_id in (6,7);
    FILE_NAME
    E:\ORADATA\DEMO111P\ABBASI_01.DBF
    E:\ORADATA\DEMO111P\ABBASI_02.DBF
    SQL>Nicolas.
    added the ROWID function usage
    Edited by: N. Gasparotto on Jun 21, 2009 11:02 AM

  • Clarification regarding Shadow Table

    Hi All,
    Clarification regarding Shadow Table. FKK_GPSHAD is a shadow table that belongs to Business Partner. The doubts that i have is
    What is Shadow Table?
    What is the purpose of Shadow table?
    Is it correct way to update Shadow table Manually(Not using any BAPi etc....)?
    Is any BAPI available to update The above mentioned Shadow table.
    Thanks in Advance..

    Reclustering InfoCubes:
    With reclustering, the InfoCube fact tables are always completely converted. The system createsshadow tables with a new clustering schema and copies all of the data from the original tables into the shadow tables. As soon as the data is copied, the system creates indexes and the original table replaces the shadow table. After the reclustering request has been successfully completed, both fact tables exist in their original state (name of shadow table) as well as in their modified state with the new clustering schema (name of original table).
    You can only use reclustering for InfoCubes. Reclustering deactivates the active aggregates of the InfoCubes; they are reactivated after the conversion.
    Reclustering DataStore Objects
    Reclustering completely converts the active table of the DataStore object. The system creates a shadow table with a new clustering schema and copies all of the data from the original table into the shadow table. As soon as the data is copied, the system creates indexes and the original table replaces the shadow table. After the reclustering request has been successfully completed, both active tables exist in their original state (name of shadow table) as well as in their modified state with the new clustering schema (name of original table).
    You can only use reclustering for standard DataStore objects and DataStore objects for direct update. You cannot use reclustering for write-optimized DataStore objects. User-defined multidimensional clustering is not available for write-optimized DataStore objects.
    Pls chk this thread:
    Shadow Table?
    http://help.sap.com/saphelp_nw2004s/helpdata/en/47/5cf74153b6ca17e10000000a155106/content.htm
    Hope this helps,
    Reward points...

  • Needs Clarification Regarding Listener.ora file

    I want a clarification regarding listener.ora file I've Listener.ora file and its content look like as:
    # listener.ora Network Configuration File: C:\oracle\product\10.2.0\db_1\NETWORK\ADMIN\listener.ora
    # Generated by Oracle configuration tools.
    SID_LIST_LISTENER =
    (SID_LIST =
    (SID_DESC =
    (SID_NAME = PLSExtProc)
    (ORACLE_HOME = C:\oracle\product\10.2.0\db_1)
    (PROGRAM = extproc)
    (SID_DESC =
    (GLOBAL_DBNAME = Oracle8)
    (SID_NAME = ORCL)
    LISTENER =
    (DESCRIPTION_LIST =
    (DESCRIPTION =
    (ADDRESS = (PROTOCOL = IPC)(KEY = EXTPROC1))
    (DESCRIPTION =
    (ADDRESS = (PROTOCOL = TCP)(HOST = 10.10.10.10)(PORT = 1521))
    I can understand every entry in this file accept the following
    *(SID_DESC =*
    *(SID_NAME = PLSExtProc)*
    *(ORACLE_HOME = C:\oracle\product\10.2.0\db_1)*
    *(PROGRAM = extproc)*
    *(ADDRESS = (PROTOCOL = IPC)(KEY = EXTPROC1))*
    Although rest of entries refer to network services(Database name) and host address and protocol etc.
    what these above refere to.
    Please clarify me in this.
    Regards,
    D.Abbasi

    (SID_DESC =
    (SID_NAME = PLSExtProc)
    (ORACLE_HOME = C:\oracle\product\10.2.0\db_1)
    (PROGRAM = extproc)
    (ADDRESS = (PROTOCOL = IPC)(KEY = EXTPROC1))
    that about External Procedure, If you don't use "External Procedure" you can remove it, by the way you should avoid default configure.
    http://download.oracle.com/docs/cd/B19306_01/network.102/b14212/advcfg.htm#sthref1331
    http://download.oracle.com/docs/cd/B19306_01/network.102/b14212/advcfg.htm#NETAG0132
    http://download.oracle.com/docs/cd/B19306_01/appdev.102/b14251/adfns_extern_proc.htm#sthref1658
    Sample create External Procedure:
    http://surachartopun.com/2008/07/create-and-run-sample-external.html

  • Help needed regards the usage of STL.....with CC

    Hi All,
    I am relatively new to SOLARIS. I am trying to figure out the options for using the STL components in the project.
    The project needs to be compiled with both CC ang g++, should support both SOLARIS and LINUX systems. That's why i am going for STL components rather than using the RW-Components of Tools.h++.
    I have the following doubts regards the usage of STL.
    CC provides the -library option to link the libraries we require.
    The following is what i understood from the documentation:
    No ( -library ) option provides - default libraries included -lCstd -lCrun -lm -lw -lcx -lc+
    -library=iostream+ - libraries included -liostream -lCstd -lCrun -lm -lw -lcx -lc+
    -library=iostream,no%Crun+ - libraries included -liostream -lCstd -lm -lw -lcx -lc+
    -library=stlport4+ - libraries included -lstlport4 -lCrun -lm -lw -lcx -lc+
    -library=iostream,no%Cstd+ - Invalid combination, some header files missing [[ *iostream, sstream* ]]
    When we try to make a new project, which of the following is recommended?
    #! - Use libCstd suppplied along with solaris package.
    When solaris makes a new release, is it always guaranteed that project is compatible with new libCstd ? (Is there a backward compatibility?)
    The STL components which can be used in the project are limited. i.e we can use only those that come along with libCstd.....right?
    Might not be compatible with other c++ compilers. ( Not compatible with g++ ) Right?
    #2 - Use libCStd along with libiostream
    Can we use STL (supported by libCstd ) + Classic-iostreams and still have the backward-compatibility?
    Compatible with other C++ compilers....provided care has been taken of the CC STL Specializations. ( Compatible with g++ ) Right?
    #3 - Use stlport4. Is it stable and backward-compatible ?
    We can exploit usage of STL to the maximum.
    Is it guranteed that the project (using -library=stlport4 ) will be backward compatible ?
    Can the SunStudio (ORACLE) organization gurantee that stlport will take care of the changes in the CPP standards ?
    i.e Is it guranteed that STLPORT and SUN-STUDIO packages will always be in sync?
    Among the above three which is preferred method to go ensuring stability and backward-compatibility.
    Thanks in advance.
    Cheers,
    Sreekar
    Edited by: 855323 on 20-Oct-2011 04:04
    Edited by: 855323 on 20-Oct-2011 04:04
    Edited by: 855323 on 20-Oct-2011 04:06

    In general, you don't need any options to use the C++ Standard Library (which includes what is sometimes loosely called the "STL"). Consider this toy program:
    // file vec.cc
    #include <vector>
    #include <iostream>
    int main()
        std::vector<int> vi(10);
        vi[1] = 1;
        std::cout << "vi[1]=" << vi[1] << '\n';
    }You can compile and run the program as either
    CC  vec.cc && a.out
    g++ vec.cc && a.outWith CC, by default you get the original libCstd. You use the -library option to select STLport or (on or selected versions of Solaris) Apache stdcxx instead. With g++, you just get the g++ library libstdc++, which should be suitable for all purposes.
    For a discussion of which library to select with CC, see this thread:
    Differnce between LibCstd and LibStlport
    The optional libiostream is provided to provide support for programs written for the very old version of iostreams that was provided with the original AT&T Cfront compiler in the 1980's and early 1990's. There is usually no reason to use this obsolete library with code written after 1998.
    The Solaris libCstd.so.1 (and libCrun.so.1) is always compatible with various releases of Studio, with one caveat: Each Studio release specifies the minimum patch level of Solaris libraries that is required to run programs created by the compiler. Any newer version of the library is guaranteed to be compatible. Thus, it is always safe to update the library, and an update might be required when using binaries created by a newer compiler.
    A version of the STLport library ships with the compiler. C++ binaries created by an older compiler that link to libstlport.so.1 should still work when linked into a program created by a newer compiler. Use of the static libstlport.a is generally not safe when mixing binary code from different compiler releases.
    Binaries created by Studio CC in default mode are not compatible with binaries created by g++. Even if you can get a program to link, which is doubtful, the program is not likely to run correctly.
    Studio CC does provide a g++ compatibility mode as follows:
    On supported versions of Linux with Studio 12.2 (C++ 5.11).
    On Solaris/x86 and supported versions of Linux with Studio 12.3 (C++ 5.12).
    In this mode, CC uses the g++ headers and the g++ runtime libraries.
    Refer to the -compat=g option in the C++ Users Guide for details.
    We plan to support the new C++ Standard, C++11, in a future compiler release. Because the ABI (binary interface) used by the current compilers is not adequate to support all the new features in C++11, we expect binaries built in C++11 mode to be incompatible with binaries created by earlier compilers. None of the existing C++ support libraries will be used in C++11 mode. A new library that provides full C++11 support will be used instead. We expect the new compiler to continue to provide the current C++03 mode as an option, being source and binary compatible with our earlier compilers.
    Edited by: Steve_Clamage on Oct 20, 2011 10:00 AM
    Edited by: Steve_Clamage on Oct 20, 2011 1:29 PM

  • Some clarifications regarding Aironet settings

    Hi,
    i need some clarifications regarding configuring Aironet stand-alone AP (in this case AIR-LAP1131AG).
    Under Security->SSID Manager:
    what is the purpose of Network ID?
    Under Guest Mode/Infrastructure SSID Settings - what is the purpose of Set Infrastructure SSID?
    and Force Infrastructure Devices to associate only to this SSID?
    Cheers,

    Assign a Service Set Identifier (SSID) to each VLAN configured on the AP. SSIDs enable endpoints to select the wireless VLAN they will use for sending and receiving traffic. These wireless VLANs and SSIDs map to wired VLANs. For voice endpoints, this mapping ensures priority queuing treatment and access to the voice VLAN on the wired network
    For further information click this link,
    http://www.cisco.com/en/US/docs/voice_ip_comm/cucm/srnd/4x/42nstrct.html#wp1098806

  • Need clarification regarding select query

    Hi,
    I need a little clarification regrding a Select senario
    I want to select data from table which have been minupulated between a certian date like between 1-DEC-10 to 31-DEC-10 and note that table does not have any time/date column. I've applied the following query to do this.
    select * from TABLE_NAME where sysdate between to_date('01-DEC-10') AND to_date('31-DEC-10');
    Would it work fine because I've tried it against a table and it returned me nothing however DML occur between time period.
    Regards,
    Abbasi

    Abbasi wrote:
    Hi,
    I need a little clarification regrding a Select senario
    I want to select data from table which have been minupulated between a certian date like between 1-DEC-10 to 31-DEC-10 and note that table does not have any time/date column. I've applied the following query to do this.
    select * from TABLE_NAME where sysdate between to_date('01-DEC-10') AND to_date('31-DEC-10');
    Would it work fine because I've tried it against a table and it returned me nothing however DML occur between time period.
    Regards,
    AbbasiAFAIK without log mining and auditing this is not possible.
    http://download.oracle.com/docs/cd/B19306_01/server.102/b14215/logminer.htm

  • Regarding the usage of HR_INFOTYPE_OPERATIONS function module in ECM badi i

    Hi All,
    We have requirement where we need to update the infotype 0015 while implementing the badi 'HRECM00_ACTIVATION'.
    To achieve this functionality we are calling the HR_INFOTYPE_OPERATION function module to update the infotype 15. But we are getting the runtime error.
    Can some one suggest how to solve this issue.
    Thanks & Regards,
    Sreelatha Gullapalli.

    thanx for ur reply...
    But i want to use Transmission Medium 5 only as i have the requirement to Send the data as an EMAIL or FAX depending on the Communication Method specified in the Partner Number.
    Actually SAP Help also explains the usage of this medium using the same requirement as mine... but then it never anywhere specified on how the Print Program will be...
    I also checked the standard output type CPAC used in Application V7 to which transmission medium 5 is assigned. And the print program assigned to this output type has no coding on sending the email using the FM. So i am very very confused.... I have seen few postings in this forum on this topic.. but not clear enuf to clarify my confusion...
    Kind Regards,
    Tanuja

  • Clarifications regarding Logistics Invoice Verification

    Hi SAP Gurus,
    I just wanted to confirm with you something that I noticed regarding the behavior of SAP MM LIV:
    1. It seems that for purchase orders with Account Assignment, you can change the Account Assignment Category in the LIV process. Say for example, I had a PO line item charged to a Cost Center; however, upon LIV, I can change the charging to say an Asset instead. I was just wondering why would SAP allow such transaction/s to happen? Wouldn't it compromise the integrity of the whole document trail (PO account assignment is different from the LIV)?
    2. Is there a way to edit the Account Assignment for a Purchase Order line item even if that item has been fully received and invoiced? I checked with the standard test system that I have at the moment and I observed that you can still edit the other PO details (quantity, price) even though the PO line item has been fully GR'ed and LIV'ed. However, the Account Assignment field is set to Display only. I checked the field selection for purchase orders and found no connection (the Account Assignment field is set to optional for all relevant factors)
    Hope to hear from you soon!
    Best regards,
    DeLo

    Hi Delo,
    Please find answers to your queries below in bold.
    1. I was just wondering why would SAP allow such transaction/s to happen? Wouldn't it compromise the integrity of the whole document trail (PO account assignment is different from the LIV)?  -
    There could be some business scenarios in some organisations, where in Procurement team uses generic cost centre (header), which is specific to particular department or group, to procure. But at the time of Invoice Receipt, AP team knows exact cost centre or order, etc  to post the amounts. To cover these kind of scenarios, SAP probably has provided the option to change acount assignment.
    2. Is there a way to edit the Account Assignment for a Purchase Order line item even if that item has been fully received and invoiced? -
    These field changes at various levels can be configured via t-code OME9.
    a) IR Changeable - This field if selected will allow to change the account assignment of a purchase order item at the time of invoice receipt.
    b) Acct. assg. changeable - This field is selected will allow to change account assignment for an item following goods receipt or invoice receipt.         
    Hope this clarifies your queries.
    -Ravi

  • Clarification Regarding Asset Accounting Chart of Depreciation Copy

    Hi,
    We have a company code called ACT
    For that we have a chart of depreciation assigned is ACT
    Number Ranges are 1000000-1999999 (For Example)
    Now we configured one more company code called RNC by Copying ACT company code and other settings from ACT company code only.
    Clarification required - In Asset Accounting for Company code RNC we are plannint to assign chart of depreciation ACT as it is.
    If we do like that - System will take the number ranges from the above ACT number ranges 1000000-1999999 ? or else system will take 1000000-1999999  number ranges from RNC company code? This is my doubt?
    I think system will share the 1000000-1999999  number ranges in Asset accounting for both company codes ACT & RNC as we are using Same ACT chart of depreciation for Both company codes.
    If we assign Same chart of Depreciation to both ACT and RNC company codes - Is it necessary to do any other configuration or no need to do any?
    Kindly clarify my doubt
    Thanks
    Supriya

    Hello,
    Make sure that the both the company codes are in a single country before assigning the same chart of depreciation.
    The number range of assets are defined at company code level. Meaning that you can have the same asset number in different company codes. The system will not share the number ranges across company codes.
    Regards,
    Ravi

  • Clarification regarding GO LIVE

    Dear all,
    I need a Clarification required regarding GO LIVE PRODUCTION server
    The client wants to upload available opening data in SAP and work parallel in both existing ERP and SAP as well.
    At the end of 2nd Qtr, They want to compare data in both the systems and want to reset the data in SAP for any differences in data.
    For example we can reset the stock through physical inventory process.
    Like wise can we reset the G/L account balances in SAP so that the client can continue only with SAP there after?
    Balance sheet will be taken in existing system for 1st 6 months and in SAP for next 6 months.
    Is this practice RECOMMENDED or do we have to set up new production client for the next six months?
    Please suggest.
    Thanks & Regards,
    AR

    Hi,
    The client needs a benchmark to understand the transactions in SAP goes well.
    At the end of the Quarter they would take balance sheet in both the SAP system and the Legacy system and compare them.
    If both of these match, thats great..
    If there are any differences, they would then need to understand why these differences are.
    Inorder to balance both of these, they would need to pass journal entries into the respective ledgers..
    The main problem would be for them to post the documents in both of the systems parallely.
    Hope your doubt is cleared.
    Cheers
    Raghu

  • Regarding Azure Usage Cost optimization required

    Hi
        This post is regarding how to reduce my Azure usage costs.We are using Microsoft Azure Cloud Services (Paas).We are  using Cloud services (Web/Worker Roles).
    So for every client specific.We are deploying in 2 instances of Web Role for webapplication and 2 Instances of Service Role for WebServices and 2 instances of Worker Role for running some jobs to meet Azure SLA for 99.95% availablilty.
    For every Client specific i am deploying the same way as above apart from test setup.So the azure costs are increasing.
    We are using this Microsoft Azure for more than 2 years.
    Due to the costs increasing recently due to added instances.we are planning on looking for options to reduce the costs.
    Please provide me some options to reduce the costs.
    Q1: Is Cloud Services (Paas) is a good option for smaller firms? or are there any alternatives?
    Q2: What options i have to optimize costs using currently using Cloud Services (PaaS) ?
    Q3: Currently my Webapp is providing like this client1.cloudapp.net, client2.cloudapp.net..etc?Is there any way i can combine all Clients into one Webapp and deploy it in 2 instances and provide some thing like mainsite.cloudapp.net/client1, mainsite.cloudapp.net/client2...
    etc..?
    Q4: Will there be any way i can try to combine all client specific Webservices into one. and connect to client specific SQL Azure databases?
    Q5:Similarly for running scheduled jobs using worker role for client specific? can i combine all into one and connect to client specific SQL Azure database to run the jobs?
    Q6: As of now i am using Azure cloud services (Web/Worker Roles) will there be any good to go for other alternatives Azure Websites / Azure VM? If yes will it help for cost optimizations even client increases?
    I had gone through some web sites / forms but did not find any good optimized solution.
    Any options for cost reduction will be highly appreciated?

    Hi PVL_Narayana,
    This is a very complex question, but in the end it all concludes with application design (architecture). Of course you can design your web and worker roles to serve all your customers in a multi-tenant approach and you can also do the same for your databases.
    However, from my expertise, there are certain scenarios where a one-bucket-satisfies-everyone approach doesn't work for those customers who want to pay extra. It's totally understandable for you to look to cutting costs considering your user pool has become
    larger. Thanks to Microsoft though, which continously drops prices to Azure, this situation usually arrives later than originally expected.
    There are some patterns I've blogged about and I spoke about at various conferences which you could benefit from in your multi-tenant application, such as the priority-queue pattern, the valet key, the event-based triggering pattern, the CQRS pattern and
    especially the sharding pattern. I'm happy to chat with you more on this subject.
    As for moving away from cloud services to Azure Websitesor VMs: I would personally not suggest that, considering that: 1. you're happy with the no-manage-required approach. Otherwise, you'll end up in updating and patching your VM's OS regularry and rebooting
    every once in a while. Using PaaS, Microsoft does that for you. 2. You don't really want to mix your worker role resources with your web resources. If your pool of users has enlarged, this probably means that your services have to satisfy more requests / sec
    as well. Therefore, allocating your web server's resources for running WebJobs rather than have specific servers which just run the worker roles could mess up your code's HW requirements and you could end up with huge performance impacts.
    Of course, these answers are quite generic given the lack of insights into your application.
    I'm happy to chat with you more on this subject, should you require any more intel.
    Alex

  • Regarding Rman

    hi all,
    Please tell me the difference between the below two commands,
    1)backup database plus archivelog delete input;
    2)backup database plus archivelog all delete input;
    and what does this "delete input" will do,
    please help me out in this.
    Regards,
    Vamsi.

    Hi Vamsi;
    To back up archived logs, datafile copies, or backup sets, then delete the input files from the operating system after the successful completion of the backup. RMAN also deletes and updates repository records for the deleted input files.
    If you specify DELETE INPUT (without ALL), then RMAN deletes only the copy that it backs up. If you specify ALL, then RMAN deletes all copies of the specified logs that it finds in the V$ARCHIVED_LOG view.
    Source:
    Summary of RMAN Deletion Methods
    http://download.oracle.com/docs/cd/B10501_01/server.920/a96566/rcmconc3.htm#462766
    Regard
    Helios

  • Just needed some clarification regarding the Viewer Builder and actually publishing your App...

    If someone could let me know if my understanding is correct, that'd be a huge help... So I've designed my publication in InDesign and exported the .zip file from the Folio Producer. I've created all of my certificates/splash screens/icons. Lastly, I just recently went through the steps of the Viewer Builder. I'm now at the stage of this process that requires me to purchase the $395 single edition so that I can enter the serial number in the last stage of the Viewer Builder. Now, to my knowledge, once I get the serial number, Viewer Builder will then give me access to an .ipa file and a .zip file. The .ipa file is for me to test on my iPad, and the .zip would be used to distribute to the App Store. I guess this is where I get confused... Let's say after I test the .ipa on my iPad, I don't like some part of my publication. I know how to update my own documents obviously, and I understand that I would have to export another .zip file from the Folio Producer, in turn requiring me to edit the exported folio link in the Viewer Builder. If I had to do that, would I need to purchase another single edition serial number since the original App was edited? Or would the same serial number apply since I'm editing that same App in the Viewer Builder? My next question is somewhat similar. Let's say all of the information is up to date and I go ahead and publish the App to the App Store. However, maybe a month later or some time in the future, I needed to update a phone number or email address--some little detail like that. Again, I understand that I'd have to update the export link in the Viewer Builder, but would I then need to create a new app since my app was already published? Would I then have to purchase another $395 single edition serial number just so that I can update my information? This seems to be the only thing in this whole process that I could use some clarification on so that I don't run into any surprises in the future. Any help would be great, thanks!

    Hi Joshua,
    When you have purchased the serial, you can rebuild your app with your updated content, as long as you use the same bundleID (applicationID), that is tied to your Apple mobile provisioning profile. The serial number is valid for a one year period.
    After you have submitted your app to Apple and it has been approved, please read: http://forums.adobe.com/message/4172167#4172167
    With kind regards,
    Klaasjan Tukker
    Adobe Systems

  • Regarding Aggregates usage

    Hello experts
    I have created a aggregate on a infocube,(using proposal from query) but how do i check whether the
    aggregate is being used or not ??
    when i check the usage tab in the 'maintaince of aggregate ' it displays 0.
    Rgds
    Priyanka

    Hi
    If you have created an aggregate for an InfoCube, activated it and entered data for it, the OLAP processor can access these aggregates automatically when the query is being executed.
    If the aggregates are being used, you can see the Usage counter in aggregate maintenance incremented as and when the respective aggregate is used.
    Also check the Help link below for more details
    http://help.sap.com/saphelp_nw70/helpdata/EN/7d/eb683cc5e8ca68e10000000a114084/frameset.htm
    Hope this helps
    Regards
    Gaurav

Maybe you are looking for