Difference between initialization with data transfer?

Dear Ones,
What is the Difference between initialization with data transfer,initialization without data transfer and early delta initialization?
Please explain me in senario basis..
Thanks
Shiva

Hello Shivananda,
See oss note 505700 :
The early delta initialization is used to admit document postings
            in the OLTP system as early as possible during the initialization
            procedure.
Regards,
Fred

Similar Messages

  • Error while executing Initialize with Data Transfer for 0FI_GL_10

    Hello All,
    Post Pre prod refresh, our timestamp for the datasource 0FI_GL_10 got reset. Due to which our deltas did not bring any records to the BW system.
    First we did an Initialize without data transfer for all the BW related datasources. The deltas were then set properly for all datasources except 0FI_GL_10.
    We then raised a message to SAP and they suggested to run the 'Initialize with data transfer' for 0FI_GL_10 so that the timestamp is set and accordingly the deltas are fixed.
    The issue now is we are getting the following error message while running INIT with data transfer.
    Job terminated in source system --> Request set to red
    Message no. RSM078
    We have copied the data till July 1st week of 2011.
    Please advice. The issue is very critical.
    Thanks & Regards
    Sneha

    Hi Arvind
    Thanks for your inputs.
    Please find below the details of the short dump.
    Runtime Errors         DBIF_RSQL_SQL_ERROR
    Exception              CX_SY_OPEN_SQL_DB
    Date and Time          09/07/2011 11:25:32
    Short text
         SQL error in the database when accessing a table.
    What can you do?
         Note which actions and input led to the error.
         For further help in handling the problem, contact your SAP administrator
         You can use the ABAP dump analysis transaction ST22 to view and manage
         termination messages, in particular for long term reference.
    How to correct the error
         Database error text........: "ORA-01652: unable to extend temp segment by 128
          in tablespace PSAPTEMP"
         Internal call code.........: "[RSQL/FTCH/FAGLFLEXT ]"
         Please check the entries in the system log (Transaction SM21).
         If the error occures in a non-modified SAP program, you may be able to
         find an interim solution in an SAP Note.
         If you have access to SAP Notes, carry out a search with the following
         keywords:
         "DBIF_RSQL_SQL_ERROR" "CX_SY_OPEN_SQL_DB"
    Information on where terminated
        Termination occurred in the ABAP program "GP_GLX_FAGLFLEXT" - in
         "FETCH_TO_ISTRUCTURE".
        The main program was "SBIE0001 ".
        In the source code you have the termination point in line 903
        of the (Include) program "GP_GLX_FAGLFLEXT".
        The program "GP_GLX_FAGLFLEXT" was started as a background job.
        Job Name....... "BIREQU_4N3PZQ12IA0X0PYGEA85IG39S"
        Job Initiator.. "BIWREMOTE"
        Job Number..... 11203300
        The termination is caused because exception "CX_SY_OPEN_SQL_DB" occurred in
        procedure "FETCH_TO_ISTRUCTURE" "(FORM)", but it was neither handled locally
         nor declared
        in the RAISING clause of its signature.
        The procedure is in program "GP_GLX_FAGLFLEXT "; its source code begins in line
        840 of the (Include program "GP_GLX_FAGLFLEXT ".

  • Diff between init with data transfer and repair full request

    hi,
    i have observed that even in the new flow we are doing init without data transfer and then repair full request
    if i do init with data transfer also i can achieve the same?
    i want to know why we need to do this ,do we have any advantage of doing init without transfer and repair full request?
    please suggest me

    Hi Venkat,
    A repair full request is loaded in cases where you get erroneous records or where there are missing records which you want to load. In such cases, a repair full request for the selective records can be loaded and is the most efficient way to correct the data, since you won't be required to load the entire data once again. Also you acheive the desired result without disturbing the delta. (You are not required to do an init w/o data transfer after a repair full, people just do it a a precaution)
    However, repair full requests should only be loaded to infoproviders with update mode 'overwrite'. Otherwise, if the InfoProvider, its very probable that you might double the value of the key-figures due to rows being added twice - in case of an InfoProvider with update mode 'Additive'. So, if your InfoProvider is additive, you will need to delete the entire data and do an 'init with data transfer' to avoid corrupting the data. (Note: you can do a repair full request for an additive infoprovider in case of lost records or if you can delete erroneous records with selective deletion.But you have to be careful with the selections lest you inadvertently load other records than required and corrupt the data)

  • I have problem with data transfer between Windows Server 2012RT and Windows7 (no more than 14kbps) while between Windows Sever 2012RT and Windows8.1 speed is ok.

    I have problem with data transfer between Windows Server 2012RT and Windows7 (no more than 14kbps) while between Windows Sever 2012RT and Windows8.1 speed is ok.

    Hi,
    Regarding the issue here, please take a look at the below links to see if they could help:
    Slow data transfer speed in Windows 7 or in Windows Server 2008 R2
    And a blog here:
    Windows Server 2012 slow network/SMB/CIFS problem
    Hope this may help
    Best regards
    Michael
    If you have any feedback on our support, please click
    here.
    Please remember to click “Mark as Answer” on the post that helps you, and to click “Unmark as Answer” if a marked post does not actually answer your question. This can be beneficial to other community members reading the thread.

  • Initialize Delta Process with Data Transfer - Split the job in two/three

    Hello,
    I am trying to load master data into BI (DEV) and there are 700,000 master data records in my source system. When I tried to load Initialize Delta Process with data transfer, at some point (after 300,000 records) my load failed in source system and job log I found that this issue "ABAP/4 processor: TSV_TNEW_OCCURS_NO_ROLL_MEMORY"
    When I informed this issue to BASIS consultant, they suggest me to split the job in 3 (Three). My question: how can I split the job in three? I can do one Initialize Delta Process load and followed by delta load.
    Regards,
    Md

    Hi Md.
    Even at Infopackage level you can reduce the number of records per package.
    Go to Infopackage Schedular Menu -> DataS. Default Data Transfer. you can reduce the Maximum size of a data packet in kByte.
    Regards,
    Pratap Sone

  • Difference between using new Date() and new Date(System.currentTimeMillis()

    Hi All,
    I have seen many open source api's where they are updating dates with the following code
    new Date(System.currentTimeMillis())
    when i print the new Date() it prints the same
    what is the difference between using new Date() and new Date(System.currentTimeMillis()) ??
    Thanks,
    J.Kathir

    when i print the new Date() it prints the same
    It does because of backward compatability. This constructor exists in version 1.3.1 but not in newer versions.Really ? Please point me to where you read that.
    API doc 1.5 : new Date() (still there, not even deprecated)
    A quick look at the code for this constructor:    /**
         * Allocates a <code>Date</code> object and initializes it so that
         * it represents the time at which it was allocated, measured to the
         * nearest millisecond.
         * @see     java.lang.System#currentTimeMillis()
        public Date() {
            this(System.currentTimeMillis());
        }

  • Disk Utility: Differences between "Zero Out Data" and "7-Pass Erase"?

    I'm wondering if anyone knows if there's a significant difference between the "Zero Out Data" erase option in Disk Utility (specifically Disk Utility 10.5.5), and the "7-Pass Erase" and "35-Pass Erase" options in same software.
    Here's why I'm asking: I have a co-worker with an iMac G5 20" 1.8GHz with 160GB internal hard drive. As a result of the power supply overheating a week ago due to dust, some hard drive problems resulted. I'm trying to assess whether these are 'soft' formatting problems that can be recovered from, or 'hard' problems requiring replacement of the hard drive and/or power supply.
    Following the failure, I removed the dust and restored the iMac to servicable form. The power supply seems to be OK now. The next thing was to attempt to recover as much data as possible from the 160GB, as the last full backup was a week old. Carbon Copy Cloner, shell copy via 'sudo cp -p -R -v', Finder copy, and DiskWarrior recovery all met with problems. TechTool Pro identified a huge swatch of unreadable sectors during repeated surface scans. Unfortunately, these unreadable sectors were located midway in the OSX boot partition (an 80GB partition), and not in the other 80GB partition devoted to lower priority video data.
    When I was satisfied I had backed up the data to the best of my abilities, I next set out to reformat the drive and see if the bad sectors could be eliminated or remapped out of existence. I did a "Zero Out Data" erasure in Disk Utility (with no errors during the erase), but TechTool Pro showed the bad sectors persisted in equal strength at the same location. I next executed a sixteen hour "7-Pass Erase" (again no errors, and confirming that it takes about an hour per 10GB). The next day when I ran TechTool pro, all of the sector errors had disappeared. I'm a bit perplexed as to why the "7-Pass Erase" seems to have recovered the use of the drive. Is it possible that there are simply thousands of bad sectors now remapped that I'm not seeing? [If so, how do I check for this?] TechTool Pro has not reported any S.M.A.R.T. issues to date on the drive. What am I to make of that?
    There are some related threads I've checked into, but I'm not sure how to properly assess my situation based on this information:
    <http://discussions.apple.com/thread.jspa?threadID=232007>
    <http://discussions.apple.com/thread.jspa?threadID=138559>
    <http://discussions.apple.com/thread.jspa?threadID=118455>
    Since the iMac has three weeks left on it's one year warranty, and I've already moved the user to another machine temporarily, I'm thinking that the smart thing to so is to send it in to Apple to have them look at the power supply and hard drive. That way, when it returns, even if there is still a lingering hardware problem, at least it will be covered under warranty for another 90 days.
    Any thoughts?
    iMac G5 20" 1.8GHz   Mac OS X (10.4.6)   1.25GB RAM, 160GB hard disk, SuperDrive

    HI, Bret.
    The only differences between "Zero Out Data", "7-Pass Erase", and "35-Pass Erase" are the number of times a binary zero is written to every bit on the disk. "Zero Out Data" writes a binary zero once, whereas the 7- and 35-Pass options write a zero seven and 35 times, respectively.
    Technically, one pass with Zero Out Data should be sufficient to map bad sectors out of service, a process also known as sparing. If a bad sector is encountered, it is both marked as "in use" in the directory's allocation table and added to the directory's "bad blocks file."
    My understanding is that the Surface Scan of Tech Tool Pro should identify bad sectors every time it is run unless the bad sectors have been locked out by the drive controller of the ATA drive itself. This is because Surface Scan checks the entire surface of the disk.
    What may have happened is that running "Zero Out Data" spared the bad blocks from a directory standpoint, but did not result in the drive's controller locking out those sectors for reasons detailed in the "Surface Scan" section of the Tech Tool Pro manual. However, the 7-Pass Erase may have resulted in the drive's controller locking out the bad sectors and why Surface Scan did not pick them up after such.
    Given the problems you described, I concur with your plan to have Apple check the affected computer. You might also want to consider purchasing an AppleCare Protection Plan for that Mac: I recommend and buy these for all my Macs.
    For some additional information on bad sectors, see the "Bad Sectors" section of my "Resolving Disk, Permission, and Cache Corruption" FAQ.
    Good luck!
    Dr. Smoke
    Author: Troubleshooting Mac® OS X
    Note: The information provided in the link(s) above is freely available. However, because I own The X Lab™, a commercial Web site to which some of these links point, the Apple Discussions Terms of Use require I include the following disclosure statement with this post:
    I may receive some form of compensation, financial or otherwise, from my recommendation or link.

  • Differences between the Portal Data Collector and the Activity Data Collect

    Hello,
    I want to know what are the differences between the Portal Data Collector and the Activity Data Collector?
    Best Regards.
    Pablo Mortera.

    All of my SQL Server instances  are sql server 2008r2 standard edition(10.50.2500). MDW is existing database, I try to setup collection sets for multiple instances and store data in one central MDW database 
    I create MDW in one instance, then run run configure Management Data Warehouse in target intance. collection sets were created successfully, but job failed with following error:
    Executed as user: COCAD\INTDEPT01SQLAgentC10. The Management Data Warehouse version "00.00.0000.00" is not supported by the current data collector. Please upgrade the Management Data Warehouse by running the Management Data Warehouse Configuration
    Wizard.  Process Exit Code 5.  The step failed.
    Thanks
    PAULqaz

  • Difference between P.O.date and P.O.created date

    Hi,
      Could you tell me difference between purchaseing doucment date and purchase order creaed date.
        Because some time in my org. po created after GR means in past date. So there is difference between
        po date and po created date. where this po created date is maintained in purchase order
               Thanks and Regards
                      Anil

    Dear Anil,
    Please check table EKKO with ta SE11.
    Created Date => EKKO-AEDAT
    Date on which the record was created
    Document date => EKKO-BEDAT
    Date on which the purchasing document was created
    EKKO-AEDAT = Date on which the record was created
    and also "Created By" also recorded automatically
    by system.
    Document date or EKKO-BEDAT normally default as
    today when you create the PO via ME21N. This field
    appear at top right of the screen. However this date
    can be change to any date that required.
    Thanks
    Loke Foong

  • Initialization without data transfer

    Hi bw experts,
    when i am loading data to ods 0fiap_o06.actually i loadded selection wise data in same infopackage for full update .after that when i try load initialization without data transfer iam getting this error
    1.Full updates already available in DSO 0FIAP_O06 ; Cannot update init./delta     RSM     98     
    2.Activation of M records from DataStore object 0FIAP_O06 terminated     RSM1 102
    after that i tried. iloaded first initialization without data transfer then i loaded full update this time these error is coming.
    1.Full 1,848 has overlapping selections with init. 0000001844 : No updates into 0FIAP_O06     RSM     94
    2.Activation of M records from DataStore object 0FIAP_O06 terminated     RSM1     102     
    right now iam doing in quality system before going to production  system ,i would like to selection wise data bcoz this is huge data.can any one suggest me how to proceed this .
    regards
    RK
    Edited by: rama krishan on May 12, 2008 12:04 PM

    Hi,
    Initially Just do Initialise Delta Process.....
    Do not select anything (Initialise w/o data transfer (or) Early delta initialisation)..
    After loading Do Delta Update
    We are successfully working on this...
    I think first do the following things..
    1. Delete data from the Cube.
    2. Load INIT
    3. After completing INIT -- Do Delta
    Now you will not have any problems..
    Hope it works...

  • Differences between operational systems data modeling and data warehouse da

    Hello Everyone,
    Can anybody help me understand the differences between operational systems data modeling and data warehouse data modeling>
    Thanks

    Hello A S!
    What you mean is the difference between modelling after normal form like in operational systems (OLTP) e. g. 3NF and modelling a InfoCube in a data warehouse (OLAP)?
    While in a OLTP you want to have data tables free of redundance and ready for transactions meaning writing and reading few records often, in an OLAP-system you need to read a lot of data for every query you do on a database. Often in an OLAP-system you aggregate these amounts of data.
    Therefore you use another principle for these database scheme. This is called star schema. This means that you have one central table (called fact table) which holds the key figures and have keys to another tables with characteristics. These other tables are called dimension tables. They hold combinations of the characteristics. Normally you design it that your dimensions are small, so the access  on the data is more efficent.
    the star scheme in SAP BI is a little more complex than explained here but it follows the same concept.
    Best regards,
    Peter

  • Difference between fully-specified data types. and generic types

    Hi,
    Can anyone tell me the difference between fully-specified data types and generic types.
    Thanks in advance.
    Regards,
    P.S.

    HI
    Generic table types
    INDEX TABLE
    For creating a generic table type with index access.
    ANY TABLE
    For creating a fully-generic table type.
    Data types defined using generic types can currently only be used for field symbols and for interface parameters in procedures . The generic type INDEX TABLEincludes standard tables and sorted tables. These are the two table types for which index access is allowed. You cannot pass hashed tables to field symbols or interface parameters defined in this way. The generic type ANY TABLE can represent any table. You can pass tables of all three types to field symbols and interface parameters defined in this way. However, these field symbols and parameters will then only allow operations that are possible for all tables, that is, index operations are not allowed.
    Fully-Specified Table Types
    STANDARD TABLE or TABLE
    For creating standard tables.
    SORTED TABLE
    For creating sorted tables.
    HASHED TABLE
    For creating hashed tables.
    Fully-specified table types determine how the system will access the entries in the table in key operations. It uses a linear search for standard tables, a binary search for sorted tables, and a search using a hash algorithm for hashed tables.
    Fully-specified table types determine how the system will access the entries in the table in key operations. It uses a linear search for standard tables, a binary search for sorted tables, and a search using a hash algorithm for hashed tables.
    see this link
    http://help.sap.com/saphelp_nw04/helpdata/en/fc/eb366d358411d1829f0000e829fbfe/content.htm
    <b>Reward if usefull</b>

  • What is the difference between initialization initiation pls explain

    **What is the difference between initialization initiation of  object pls explain with example............**
    WHAT is the default constructor..........

    886152 wrote:
    **What is the difference between initialization initiation of  object pls explain with example............**There is no such thing in the official Java vernacular as "initiation."
    WHAT is the default constructor..........[url http://java.sun.com/docs/books/jls/third_edition/html/classes.html#16823]JLS 8.8.9 Default Constructor says:
    >
    If a class contains no constructor declarations, then a default constructor that takes no parameters is automatically provided:
    * If the class being declared is the primordial class Object, then the default constructor has an empty body.
    * Otherwise, the default constructor takes no parameters and simply invokes the superclass constructor with no arguments.

  • Difference between "change auth data " and "expert mode" in pfcg

    Gurus ,
    i need to know , what is the difference between
    "change  authorization data "
    and
    "expert mode for profile generation" in PFCG .
    and what is the significance of the three options that we get when we choose expert mode ?

    Hi Susin,
    To change the authorization data for the transactions assigned to the role, choose Change Authorization Data or Expert Mode for Profile Generation.
    If you are generating the profile for the first time, there is no difference between the two modes.
    When you change a role, you must regenerate the authorization profile. In this case, the status of the profile generation is displayed red or yellow at the top of the Authorizations tab page.
    If the status display is red, you must perform an authorization data comparison, since the menu was changed since the last profile generation or no authorization data exists.
    If the display is yellow, the authorization data for the role was changed and saved after the last generation. The generated profile is no longer current. You need to regenerate it.
    The significance of the three options that we get when we choose expert mode is as follows:
    Choose one of the options to maintain the authorization values (in normal mode, the correct option is automatically set):
    1> Delete and recreate profile and authorizations
    All authorizations are recreated. Values which had previously been maintained, changed or entered manually are lost. Only the maintained values for organizational levels remain.
    2> Edit old status
    The last saved authorization data for the role is displayed. This is not useful, if transactions in the role menu have been changed.
    3> Read old status and compare with new data
    If you change transactions in the role menu, this option is the preconfigured. The profile generator compares the existing authorization data with the authorization default values for the menu transactions. If new authorizations are added during this process, they receive the status New. Authorizations that already existed receive the status Old.
    Note the following during the comparison:
    Values for organizational levels that are no longer required are deleted, all others are retained. If new organizational levels are added, you need to maintain them.
    The standard authorization for the object S_TCODE is always automatically filled with the current transactions from the role menu. It cannot be copied or manually changed, only deactivated.
    Thanks,
    Saby..

  • What is difference between tha master data and tranction data

    Hi, Gurus
    this is  sudhakar
    i want to no what is the difference between the transaction data and master data  give with examples .
    Thanks
    sudhakar

    Hi,
    The most important data classes are master data, transaction data, organizational data and system
    data.
    Master data is data that is rarely modified. An example of master data is the data of an address file, for
    example the name, address and telephone number.
    Transaction data is data that is frequently modified. An example is the material stock of a warehouse,
    which can change after each purchase order.
    Organizational data is data that is defined during customizing when the system is installed and that is
    rarely modified thereafter. The country keys are an example.
    System data is data that the R/3 System itself needs. The program sources are an example
    for more info go through..
    [http://help.sap.com/saphelp_nw70/helpdata/]
    regards,
    NR

Maybe you are looking for

  • Premiere CC 2014.1 Crash

    I keep on getting this error; I reinstalled Premiere and this keeps happening. I checked all graphic card drivers, and they're all updated. I am using Lenovo T540p with the following specs: Intel Core i7-4800MQ CPU @ 2.70 GHz 16.0 GB RAM 64-bit NVIDI

  • Please help!!! rotating a video clip!!!!

    Hi all, I went to see Robbie Williams in concert and took a video of one of the songs... problem is, like a dork, a took the video in portrait format on my camera, so now it imports in landscape with robbie singing on his side... Does anyone know how

  • Setting multiple R12 iModules on same external tier which is in DMZ Zone

    Experts, Could some one guide me on how to setup more than one iModules on same external tier which is in DMZ Zone. We have already setup iStore and its working fine. Now we would like to extend it to other modules like : iPayables, iReceivables and

  • Adobe Photoshop crashing when selecting print preferences

    Hi There, It seems that within the past 2 days in our company we have had users complain that Adobe Photoshop is crashing when they go to print and print preferences. They say a window appears and says "Photoshop has stopped working and windows is lo

  • AQ JMS: non-durable subscriptions not cleaned up

    Whenever I create a non-durable subscription I see a TSUB_... subscription in the AQ$ tables. This subscription does not disappear when the client terminates abnormally. This can be very annoying because these queues fill up quickly and deleting a bi