Business Intellience and Data Archiving

Once you have implemented Data Archiving, what are the possible ways of allowing BI standard and custome extractors to look at both SAP ECC online and archived data (the archived data sits on Open text). How do you deal with BI complete reload and refresh requirement for a specific data if the data has been already been archived?

Is there any addiotional things we need to do at info source level after activating the content? Selecting the correct data source etc.
If there is a enhancement to the extraction structure then what all steps manually  we need to do at info source level? When we replicate the data source which has been enhanced, will it create the infoobjects automatically so that they can be used in infosource communication structure?
As I am new to the system, I am running in to all these questions.
Appreciate any help in step by step how things will work.
Thanks.

Similar Messages

  • Need material on business content and data extraction

    hi
      friends
                iam new to BW i Need material on business content and data extraction  , plzz send me without help.sap material to my mail id [email protected]

    Hi,
    Pls check ur inbox....
    Also go thru this link....
    http://help.sap.com/saphelp_nw2004s/helpdata/en/37/5fb13cd0500255e10000000a114084/frameset.htm
    Hope it helps,
    Thanks,
    Happy Life,
    Aravind

  • Column Modification impact to Data foundation, Business layer and Data source in DS 1.2

    Issue
    An existing Business Layer that is consumed by a Data Source definition in Design Studio 1.2 Version: 12.0.2.201311041742 seems to invalidate the entire Data Source object after modification to the base tables DDL (data definition).  Information Design Tool used is Information Design Tool 4.1, Build:14.1.1.1036, Version: 4.1 Support Pack 1.
    Background
    A modification was needed a base tables DDL (data definition).  Specifically a column that stores the Date value as a varchar needed to update to a Date data type.  After the modification was completed the Data Foundation and Business Layer in the Information Design Tool (IDT) needed
    modification as well to reflect the updated column data type.  Refreshing the Data Foundation and/or the Business Layer did not update the column Data Type.  The step of dropping and re-inserting the updated table in the Data Foundation and subsequently the Business Layer updated
    the expected columns Data Type as expected. All the existing Data Source definitions that were in use throughout the application (Design Studio) were invalidated causing the following error after the DDL update.
    After rebuilding the Data Source definition in Design Studio, the issue was no longer there however, the existing Data Sources were still non-functional and were subsequently deleted.
    In Sum
    Is there another way to make changes to the base table(s) without negatively impacting the Data Foundation (IDT)àBusiness Layer (IDT)àData Source Definition (Design Studio) to the point where Data Sources on the Design Studio have to be redone preceded with similar effort at the IDT level?  Is there a step(s) that I missed?
    Any insight would be appriciated.
    Regards,
    Kris

    Hi Kris - it looks like you have not downloaded the latest SP's for Design Studio 1.2
    http://service.sap.com/sap/support/notes/1954748
    I recall in late May SAP issues SP2, and earlier in the year they issued SP1 - could you apply both SP's, client-side and platform-side?

  • Business Content and Data Sources automatic binding

    Hi,
    I am not clear about the Data Sources and Business Content sync.
    If I transfer the datasources in SBIW(RSA5)  and replicate in BW, how it will automatically bind to the infosources and data targets when we activate the relevant infoobjects in BW Business content. As part of Business content activation, will it search for datasources and bind to it automatically? What steps are needed in order to make the everything hooked up correctly?
    Can anyone clarify to me how it works? Please explain step-by-step.
    Thanks.

    Is there any addiotional things we need to do at info source level after activating the content? Selecting the correct data source etc.
    If there is a enhancement to the extraction structure then what all steps manually  we need to do at info source level? When we replicate the data source which has been enhanced, will it create the infoobjects automatically so that they can be used in infosource communication structure?
    As I am new to the system, I am running in to all these questions.
    Appreciate any help in step by step how things will work.
    Thanks.

  • Business Objects and Data Services on VMWare

    My company has a requirement to deploy Business Objects 4.0 as well as Data Services.
    We are going to be deploying them on Windows Server 2008 R2, hosted on a VMWare cluster.  I have reviewed the PAM, but my question is:
    Will these run using SQL Server 2008 R2 as a database? I see "SQL Server 2008" and I am just curious whether R2 is also supported?
    Thanks.

    Hi,
    SQL Server 2008 R2 is listed in the PAM for DS as Repo DB. For BI4 it is also working and supported. I installed it already on it.
    Regards
    -Seb.

  • Business Graphics and Data Tables

    Hi All,
    I have a SAPGUI application which uses the SAP Chart to display a pretty simple column chart with a data table underneath.
    With the SAPGUI Chart Options this is set as so: [Chart Options |http://www.assetcentric.com.au/images/chart-options.png]
    The result is a chart looking like so:
    [Chart Example|http://www.assetcentric.com.au/images/chart-example.png]
    Does anybody know whether this is possible to achieve with the IGS from WebDynpro?
    Regards
    Andrew

    Hi,
    If you are using webdynpro for ABAP then we cna use the Business Graphics UI element where we have differenet chart optins( bar, pie, lines etc) to have the graphs.
    BG in Webdynpro -
    http://help.sap.com/saphelp_nw70/helpdata/EN/ed/258841a79f1609e10000000a155106/content.htm
    For WD4Java we need the separate IGS  server to render the graphs, but in WD4A we dont need it. we can design webdynpro applicaitons in SE80 and can run the applciations in browser direclty.
    we can use the data from BAPI by making a service call to that BAPI, and bind the tables to this BG UI element to display.
    Regards
    Lekha

  • Business Rules and Dates

    I have many Java POJOs imported as facts. In each of these POJOs there is at least 1 date property.
    How am I to use these dates in my rules. for example:
    pojo.date > '08/22/06'
    pojo.date < today
    println(pojo.date)
    Although I have imported the java date class and have been over the docs many times, I cannot seem to figure out how to use the dates properly.
    I'm also having a similar problem where I have large numbers stored as doubles. When I make a rule like: pojo.largenumber > 3000 I get the error :
    Cannot perform operation. 'RUL-01726: Data type operator for >, at position 0: java.lang.Long > int '
    Thanks for the help,

    smartel2 wrote:
    I understand we need to compare like values, just as we would in java and other
    languages. however, I am trying to use the Rule Author to enter my rules, as my end
    user / business analysts would. When I use the rule author I can't seem to use any of
    the Date methods for example using the expert mode in the Rule Author
    pojo1.getStarttime ().compareTo(pojo2.starttime)
    gives: "Cannot perform operation. 'RUL-01717: "compareTo" is not a valid identifier
    name. '" when I apply it.Ah, I see. You are doing this correctly. You need to import the Date class into the datamodel and then check the "expand" option for the getStarttime() method of the fact type of pojo1 (or, the expand option if there is a starttime bean property). It should show up then.
    I also don't understand why my previous example "pojo.largenumber > 3000" would
    need a conversion. Shouldn't the interpreter be smart enough to know that the left
    side of the expression is a double and auto cast the right side constant '3000' as a
    double to satisfy the condition?The error indicated that "largenumber" was a Long object, not a primitive double. If largenumber was a double, the cast should be done correctly. If it's not, it may be a bug. Post the code for the pojo class if you think otherwise.
    How do I use the Rule Author to create these kinds of rules?I think the above mostly covers this one.
    How would I use the Rule Author to create a rule for "pojo.startdate is after Sept 1, 2005"One suggestion I have is that you use the Calendar class instead of Date for doing date manipulation. Date is mostly deprecated, and basically only servers as a wrapper for a long number (seconds since the epoch). Most importantly, Date is not internationalizable.
    Calendar has full support for locales, timezones, and all of the other necessary stuff for appropriately handling dates. A simple example of the reason for this class is that date formats differ significantly, for instance, "01/07/89" is January 7th, 1989 to an American, but is July 1st, 1989 to a European. Depending on the format, it could be July 1st, 0089 to a computer.
    So, the answer to you question is to create a final variable that represents the date (either a Calendar or Date) and then use that. With Calendar, you have to also use the SimpleDateFormat object to initialize it. I would probably write a function to initialize all of the global final date objects.
    How do I use standard java objects and there methods in the Rule Author, like Date.Answered above I think, but the short is to import them into the DataModel and then check the "expand" option for the property or method of that type that you want to access its properties and methods.
    Do I have to create all my rules in pure RL language (full script) ?Could you clarify this question?
    And finally, fully integrated date support is one of the things that's being worked on for the 11g release, so this should be much easier to handle in the future.

  • Data Archiving and ABAP

    Hi All,
    I am an ABAPer and recently came accross the process of data archiving..
    Can anyone tell me what significance data archiving has w.r.t ABAP.Are they interrelated at any point??
    Waiting for Reply..
    Shilpa

    Hi Shilpa,
    As a developer you must be aware of the fact that what ever application are provided by SAP they all are using codes written in ABAP in background .
    Similarly we have a tool Archive Developent Kit (ADK) provided by SAP that uses various programs written in ABAP grouped in the form of Archive object to perform Archiving Successfully in any system.
    Yes Archiving and ABAP both are interrelated and an ABPER can very well understand how actually there programs functions at runtime while we Archive data from the database.
    Another imporeant thing is apart from the standard Archive Objects provided by SAP sometimes  requiremnets comes for  custom Objects to be created which requires good indepth knowledge of ABAP.
    I hope my answer will help you understand the assosiation of ABAP and Data Archiving.
    For more you can go through the below link ::
    http://help.sap.com/saphelp_47x200/helpdata/en/8d/3e4d22462a11d189000000e8323d3a/frameset.htm
    -Supriya

  • Business Partner Data Archiving - Help Required

    Hi,
    Am New to Data Archiving and need help to Archive Business Partner in CRM. I tried to Archive some BP's but it was not archiving. Kindly throw some light on it.
    Problem we face are:
    1) When we try to write a Business Partner to an Archive File the Job Log Shows finished but no files are created in the System.
    2) When we try to delete the BP from Data Base it doesn't show the Archived BP's that could be deleted (I guess this is because there are no archived file).
    Archive Object Used is CA_BUPA. We have created a Variant and set the Time as Immediate and Spool to LOCL.
    Kindly let us know if there is any step we are missing here and if we are on the wrong track.
    All answers are appreciated.
    Thanks,
    Prashanth
    Message was edited by:
            Prashanth KR

    Hi,
    To archive a BP following steps are to be followed.
    A. Mark the BP to be archived in transaction BP >status tab
    B. Run dacheck
    FYI : Steps on how to perform dacheck :
    1.  In transaction DACONTROL change the following parameter to the
    values : for CA_BUPA .
    No. calls/packages   1
    Number of Objects   50
    Package SizeRFC     20
    Number of Days       1 (if you use mobile client it should be more)
    2.If a business partner should be archived directly after the
      archiving note was set, this can be done by reseting the check
       date with report CRM_BUPA_ARCHIVING_FLAGS.
       here only check (set) the options :
       - Archiving Flag
       - Reset Check Date
      Reset all options should be unchecked.
    3. go to dacheck and run the process
    4. This will change the status of the BPs to 'Archivable'
       *Only those bp's which are not involved in any business transactions,
        install base, Product Partner Range (PPR) will be set to archivable
    The BP's with status 'Archivable' will be used by the archiving
         run.
    Kindly refer note 646425 before goint to step C ***
    C.Then run transaction CRM_BUPA_ARC.
    - Make sure that the "selection program" in transaction "DACONTROL"
       is maintained as " CRM_BUPA_DA_SELECTIONS"
      Also create a variant, which can done by excecuting
      CRM_BUPA_DA_SELECTION and enter the variants name for the field
      Variant. This variant is the buisness partner range.
    - Also please check note 554460.
    Regards, Gerhard

  • Data archiving and data cleansing

    hi experts,
    Can any one tell me step by step guide for Data archiving and Data cleansing of SAP-ISU Object.
    what is the difference between  Data archiving and Data cleansing .
    Thanx & Rgds

    Data Archiving: So many objects are there you can look some of them..........
    ISU_BBP IS-U Archiving: Budget Billing Plan
    ISU_BCONT Business Partner Contacts (Contract A/R + A/P)
    ISU_BILL IS-U Archiving: Billing Document Header
    ISU_BILLZ IS-U Archiving: Billing Line Item
    ISU_EABL IS-U Archiving: Meter Reading Results
    ISU_EORDER IS-U Archiving: Waste Disposal Order
    ISU_EUFASS Archiving of Usage Factors
    ISU_FACTS Installation Facts
    ISU_INSPEC IS-U Archiving: Campaigns for Inspection List
    ISU_PPM Prepayment Documents
    ISU_PRDOCH IS-U Archiving: Print Document Header
    ISU_PRDOCL IS-U Archiving: Print Document Line Item
    ISU_PROFV IS-U Archiving: EDM Profile Values
    ISU_ROUTE IS-U Archiving: Waste Disposal Route
    ISU_SETTLB Settlement Document
    ISU_SWTDOC Archive Switch Document
    Go to SARA t-code and enter the object CA_BUPA for business partner then Press F6 you will get  step by step documentation. please follow the procedure for all the objects.
    Regards,
    Siva

  • Put Together A Data Archiving Strategy And Execute It Before Embarking On Sap Upgrade

    A significant amount is invested by organizations in a SAP upgrade project. However few really know that data archiving before embarking on SAP upgrade yields significant benefits not only from a cost standpoint but also due to reduction in complexity during an upgrade. This article not only describes why this is a best practice  but also details what benefits accrue to organizations as a result of data archiving before SAP upgrade. Avaali is a specialist in the area of Enterprise Information Management.  Our consultants come with significant global experience implementing projects for the worlds largest corporations.
    Archiving before Upgrade
    It is recommended to undertake archiving before upgrading your SAP system in order to reduce the volume of transaction data that is migrated to the new system. This results in shorter upgrade projects and therefore less upgrade effort and costs. More importantly production downtime and the risks associated with the upgrade will be significantly reduced. Storage cost is another important consideration: database size typically increases by 5% to 10% with each new SAP software release – and by as much as 30% if a Unicode conversion is required. Archiving reduces the overall database size, so typically no additional storage costs are incurred when upgrading.
    It is also important to ensure that data in the SAP system is cleaned before your embark on an upgrade. Most organizations tend to accumulate messy and unwanted data such as old material codes, technical data and subsequent posting data. Cleaning your data beforehand smoothens the upgrade process, ensure you only have what you need in the new version and helps reduce project duration. Consider archiving or even purging if needed to achieve this. Make full use of the upgrade and enjoy a new, more powerful and leaner system with enhanced functionality that can take your business to the next level.
    Archiving also yields Long-term Cost Savings
    By implementing SAP Data Archiving before your upgrade project you will also put in place a long term Archiving Strategy and Policy that will help you generate on-going cost savings for your organization. In addition to moving data from the production SAP database to less costly storage devices, archived data is also compressed by a factor of five relative to the space it would take up in the production database. Compression dramatically reduces space consumption on the archive storage media and based on average customer experience, can reduce hardware requirements by as much as 80% or 90%. In addition, backup time, administration time and associated costs are cut in half. Storing data on less costly long-term storage media reduces total cost of ownership while providing users with full, transparent access to archived information.

    Maybe this article can help; it uses XML for structural change flexiblity: http://www.oracle.com/technetwork/oramag/2006/06-jul/o46xml-097640.html

  • Delete the Business Entity and the Building master data

    Dear all ,
    Please I need to know the way to delete the Business Entity and the Building master data from the system.

    Hi Khairy,
    You can allow the deletion of the object(s) at the below mentioned path.
    RECACUST -> Tools -> Deletion of Real Estate Objects Without Archiving.
    You can go to the relevant master data transaction and get an option to delete in the initial screen, an example is shown below.
    Note that there are prerequisites for every type of object deletion.
    Regds,
    Hardik Sharma

  • SAP Data Archiving in R/3 4.6C and ECC 6.0.

    Hi Guys,
        Need some suggestions,
    We currently working on SAP R/3 4.6 and have plans to upgrade it to ECC 6.0,
    In the mean time there is an requirement for SAP Data Archiving for reducing the database size and increase system performance.So wanted to know if its better to do Data Archiving before upgrade or after. Technically which will be comfortable, Also wanted to know if any advanced method available in ECC 6.0 compared to SAP R/3 4.6.
    Please provide your valuable suggestion.
    Thanks and Regards
    Deepu

    Hi Deepu,
    With respect to archiving process there will be no major difference in 4.6 and ECC 6.0 system. However you may get more advantage in ECC6.0 system because of Archive routing and also upgraded Write, Delete programs (upgraded program will depend on your current program in 4.6 system). For example In 4.6 system for archive MM_EKKO write program RM06EW30, archiving of purchase document is based on company code in the selection criteria and there is no preprocessing functionality. In ECC 6.0 you can archive by purchase organization selection criteria and preprocessing functionality will additional help in archiving of PO.
    In case if you archive documents in 4.6 and later upgrade it to ECC 6.0, SAP system will assure you to retrieve the archived data.
    With this i can say that going with archiving after upgrade to ECC 6.0 system will be better with respect to archiving process.
    -Thanks,
    Ajay

  • A question on different options for data archival and compression

    Hi Experts,
    I have production database of about 5 terabyte size in production and about 50 GB in development/QA. I am on Oracle 11.2.0.3 on Linux. We have RMAN backups configured. I have a question about data archival stretegy.  To keep the OLTP database size optimum, what options can be suggested for data archival?
    1) Should each table have a archival stretegy?
    2) What is the best way to archive data - should it be sent to a seperate archival database?
    In our environment, only for about 15 tables we have the archival stretegy defined. For these tables, we copy their data every night to a seperate schema meant to store this archived data and then transfer it eventualy to a diffent archival database. For all other tables, there is no data archival stretegy.
    What are the different options and best practices about it that can be reviewed to put a practical data archival strategy in place? I will be most thankful for your inputs.
    Also what are the different data compression stregegies? For example we have about 25 tables that are read-only. Should they be compressed using the default oracle 9i basic compression? (alter table compress).
    Thanks,
    OrauserN

    You are using 11g and you can compress read only as well as read-write tables in 11g. Both read only and read write tables are the candidate for compression always. This will save space and could increase the performance but always test it before. This was not an option in 10g. Read the following docs.
    http://www.oracle.com/technetwork/database/storage/advanced-compression-whitepaper-130502.pdf
    http://www.oracle.com/technetwork/database/options/compression/faq-092157.html

  • What is data archiving and DMS(Data Management System) in SAP

    what is data archiving and DMS(Data Management System) in SAP
    Welcome to SCN. Before posting questions please search for available information here and in the web. Please also read the Rules of Engagement before further posting.
    Edited by: kishan P on Aug 31, 2010 1:06 PM

    Hi,
    Filtering at the IDoc Level
    Identify the filter object (BD59)
    Modify the distribution model
    Segment Filtering
    specify the segments to be filtered (BD56)
    The Reduced IDoc Type
    Analyze the data.
    Reduce the IDoc type (BD53)
    Thanks and regards.

Maybe you are looking for