Initial Load Strategy/Approach

Hello Gurus,
  Does any has a strategy/approach to load 10 yrs of historical data during Initial load (Initial Extract Date is 01/01/2005) using BI Apps 11.1.1.8.1?
Thanks,
RK

Years doesnt matter. What matters is the volume of data.
If it is more than 5M, better to go with bulk loading.

Similar Messages

  • MacBook Pro (mid 2010) - Slow Startup and Initial Load of Apps after Mavericks Install

    Hello everyone,
    Long time Snow Leopard user here.  I've been absolutely, 100% thrilled with my MBP's performance / speed since mid 2010.  Since installing Mavericks I've noticed two things:
    1. Initial on boot load times seem much slower
    2.  The first time you open an app (Safari, even TextEdit) is far slower than in Snow Leopard.
    Now granted, once the computer has been on for about 5 minutes, and once you've had an application open for 1-2 minutes, than performance is okay once again.  But it seems that quite literally everytime you do "anything" for the first time since the computer has been turned on you run the risk of a spinning beachball and an extended wait.  This simply never happened in Snow Leopard.
    I'm not a computer science major but I do understand that when you do something for the first time since a reboot it will need to load new files into the RAM which takes time.  However, I've been using this computer since 2010 and have never experienced this behavior / speed / delay, etc... something seems fundamentally different.
    For example, even just opening finder, and navigating to the Applications folder results in sluggish behavior the first time.  Also, anytime you first click the top menu bar in an application you run the risk of a beachball.
    It doesn't seem like a failing hard drive because I can read/write large chunks of data one an app is finally loaded.  It's just initially loading the app takes quite literally 10-20 times longer than Snow Leopard.  It seems like perhaps Mavericks itself, the core of the new OS was perhaps installed in a bad sector of the HD, but none of my other files are on bad sectors... if that makes sense.  Again - I'm not a computer science major so I'm largely just making things up
    It also seems as if multitasking and switching between several open apps has seen a decrease in performance.  Could it be that perhaps Mavericks RAM management is superior to Snow Leopard, but only if there is a TON of RAM to go around?  Could it be as simple as Macs with only  4GB installed should not be allowed to even install Mavericks? 
    Does this sound normal for Mavericks considering I have a C2D processor, a non SSD, and only 4GB ram?  I know those specs are "outdated" for 2014, but keep in mind these were the specs that I've always had and with Snow Leoapard I was consistently thrilled with the speed of the computer.
    I suppose my question is: should my strategy be to simply turn the computer on, get a cup of coffee and let everything load and maybe even have my frequently used apps load on startup, etc... and carry on as normal?  Does something seem off?  Should I invest in a new SSD / RAM upgrade?  I know the upgrades couldn't hurt, but the issue is that I was perfectly content until upgrading to Mavericks
    I've tried booting into OS X Recovery mode and using the Disk Utility.  I've tried Safe Mode.  I've included a "EtreCheck" report below.  I'm open to any ideas / suggestions.
    Thanks!!
    Hardware Information:
              MacBook Pro (13-inch, Mid 2010)
              MacBook Pro - model: MacBookPro7,1
              1 2.4 GHz Intel Core 2 Duo CPU: 2 cores
              4 GB RAM
    Video Information:
              NVIDIA GeForce 320M - VRAM: 256 MB
    System Software:
              OS X 10.9.1 (13B42) - Uptime: 0 days 0:10:48
    Disk Information:
              Hitachi HTS545025B9SA02 disk0 : (250.06 GB)
                        EFI (disk0s1) <not mounted>: 209.7 MB
                        Macintosh HD (disk0s2) / [Startup]: 249.2 GB (228.46 GB free)
                        Recovery HD (disk0s3) <not mounted>: 650 MB
              MATSHITADVD-R   UJ-898 
    USB Information:
              Apple Inc. Built-in iSight
              Apple, Inc. Keyboard Hub
                        Logitech USB Receiver
                        Apple Inc. Apple Keyboard
              Apple Internal Memory Card Reader
              Apple Inc. BRCM2046 Hub
                        Apple Inc. Bluetooth USB Host Controller
              Apple Computer, Inc. IR Receiver
              Apple Inc. Apple Internal Keyboard / Trackpad
    FireWire Information:
    Thunderbolt Information:
    Launch Daemons:
              [System] com.adobe.fpsaud.plist 3rd-Party support link
    Launch Agents:
              [System] com.adobe.AAM.Updater-1.0.plist 3rd-Party support link
              [System] com.adobe.AdobeCreativeCloud.plist 3rd-Party support link
    User Launch Agents:
              [not loaded] com.adobe.AAM.Updater-1.0.plist 3rd-Party support link
              [not loaded] com.google.keystone.agent.plist 3rd-Party support link
    User Login Items:
              None
    Internet Plug-ins:
              FlashPlayer-10.6: Version: 11.9.900.152 - SDK 10.6 3rd-Party support link
              Flash Player: Version: 11.9.900.152 - SDK 10.6 Outdated! Update
              QuickTime Plugin: Version: 7.7.3
              JavaAppletPlugin: Version: 14.9.0 - SDK 10.7 Outdated! Update
              AdobeAAMDetect: Version: AdobeAAMDetect 2.0.0.0 - SDK 10.7 3rd-Party support link
              Default Browser: Version: 537 - SDK 10.9
    Audio Plug-ins:
              BluetoothAudioPlugIn: Version: 1.0 - SDK 10.9
              AirPlay: Version: 1.9 - SDK 10.9
              AppleAVBAudio: Version: 2.0.0 - SDK 10.9
              iSightAudio: Version: 7.7.3 - SDK 10.9
    iTunes Plug-ins:
              Quartz Composer Visualizer: Version: 1.4 - SDK 10.9
    3rd Party Preference Panes:
              Flash Player  3rd-Party support link
    Old Applications:
              None
    Time Machine:
              Time Machine not configured!
    Top Processes by CPU:
                   4%          WindowServer
                   1%          EtreCheck
                   0%          mds
                   0%          mds_stores
                   0%          airportd
    Top Processes by Memory:
              111 MB          com.apple.IconServicesAgent
              94 MB          Finder
              57 MB          Dock
              53 MB          SystemUIServer
              53 MB          EtreCheck
    Virtual Memory Information:
              1.64 GB          Free RAM
              1.42 GB          Active RAM
              314 MB          Inactive RAM
              390 MB          Wired RAM
              316 MB          Page-ins
              0 B          Page-outs

    Wish I could be of some help, but I'm having similar issues on my mid 2011 macbook pro with 8 GB ram. It's very frustrating. I haven't worked on a computer this slow since the early 90's. Even simple things like signing in to this site bring up the beach ball for maybe ten seconds. When opening up apps I often have to wait minutes, and often end up force quitting when they hang. I'm using force quit a dozen or more times a day, when it used to be very rare.
    I don't know if it's Mavericks. I installed a new hard drive in November, but the problems really started getting bad since the latest upgrade.
    It's not just opening apps. If I have an app open in the background and switch to it I have similar issues. Watching a simple video is often accompanied by stalls and delays - which is really bad as I'm often using the laptop to edit video. It's become unuseable for work.
    Perhaps the most frustrating is clicking on a link, only to find Safari decides to complete a previous scroll that it had paused. The page jumps down, and sometimes I end up clicking on a link I didn't want. Then I have to wait for THAT spinning beach ball before I can go back to the previous spinning beach ball.
    It can take up to ten minutes to restart. I've tried everything I could find on these pages so far, and nothing has worked.
    I thought maybe it was something I've done (and perhaps it is), but I keep seeing others with similar issues.

  • Initial load of data to UCM for Customer Hub PIP

    Hi,
    What is the recommended approach to have XREF tables populated during an initial load to UCM (via EIM), when the Accounts already exist in Siebel CRM?
    Our approach as of now, using EIM, is as follows:
    1) Set required customer information in EIM_UCM_ORG
    2) Look up the customer's existing Row_ID in Siebel, and populate EIM_UCM_ORG.EXT_SYSTEM_NUM and EIM_UCM_ORG.UCM_EXT_ID) accordingly
    3) Run the EIM job and UCM Batch Process to import the customer into UCM
    The account then appears in UCM with the correct reference to siebel/row_id under the "external account IDs" tab. HOWEVER, it also contains a reference to a newly created duplicate record for that account in Siebel. Looking at the xref tables, there is no reference to the existing Siebel/row_id specified in the EIM batch load, and our hypothesis is that this is the reason the account cannot be found (and a duplicated is created). What we want to achieve here is to tell UCM that the accounts we are inserting do infact already exist in Siebel CRM, and can be identified by the row_id that we pass along.
    The relevant system versions are Customer Hub PIP 11g with AIA 3. Siebel CRM and Siebel UCM are on patch 8.1.1.7 (and pertinent ACRs have been incorporated in the two Siebel instances).
    Any hints or suggestions on how to approach this would be appreciated
    -M
    Edited by: 968713 on Nov 1, 2012 5:05 AM
    Edited by: 968713 on Nov 1, 2012 5:05 AM
    Edited by: 968713 on Nov 1, 2012 5:06 AM

    Do you really need to populate the XREF table/transaction History table for initial load?

  • Handling password while initial load process

    Dear Experts,
    This is about password handling in IDM. While doing initial load, I do not want to bring passwords from my target systems (AD/SAP) into IDM.
    So which password(s) the users will use to login into target systems (AD/SAP) after initial load ? What can be achieved with pass "update system privilege trigger attribute" which is available in initial load job ?
    Is it something like, IDM creates a default password on initial load which is sent back to target systems(from which initial load was done) which changes the password for the target systems to this new default password ?
    Can we handle this default password being sent to target systems with the help of this pass "update system privilege trigger attribute" in initial load? so that this default password is not sent to target systems ??
    So if the default password is not sent back to target systems after initial load, then users will keep using their existing passwords for their login in the target systems. After that, If I need to assign UMEJAVA only privilege to the users, the password for the target systems will be changed with the default password being sent on email to the users. Since the password on AD is now changed, how the users gonna login into AD to check their emails for the
    new password ?
    It seems I have written a BIG query here .... sorry for that. But please let me know if any thing above does not make any sense.
    Also please share your views/expertise/best practice on the same.
    Many thanks in advance!
    Naveen
    Version: IDM 7.2

    Hi Naveen,
    Firstly to answer your query on what basis we decide he backend type for your UMEJAVA repository, its the business. If you want users to authenticate against AD, when they try to login to IDM UI, you have to configure the LDAP as backend and you have to choose datasource as Microsoft ADS (Deep Hierarchy) + Database (ume database).
    If you want the Users to authenticate against UME database, by default ume points to UME database and you need to create the users in the UME database.
    So, if you have configured AS JAVA with ADS+Database, in IDM you have to select the repository as SAP netweaver as Java (ldap backend)
    In the repository constants, there is an attribute called BACKED_REPOSITORY which should be your AD repository name that is configured.
    If you have a look at the AS JAVA connectors in the provisionign framework, in the create user plugin, IDM first checks for backend type. If it is LDAP backend, it just sets the JAVA account attibute, If the backend type is DB, IDM will create the user in the UME database.
    Considering your system details, i would suggest you the below approach.
    1. Configure your UMEJAVA with Microsoft ADS (Deep Hierarchy) + Database (ume database). For more information on how to configure your     
        UMEJAVA with LDAP backend refer to this link
    2. So, now the users who try to login to IDM UI or any app on AS Java, will be authenticated against your active directory.
    3. Perform the initial load from HR.
    4. Perform the initial load from AD.
    5. Perform the initial load from you AS UMEJAVA.
    6. Now, all the user information/role assignment information is loaded to IDM.
    7. Now lets discuss about password management. There are two things here
      a. Change password (by user)  - User changes password in IDM --> password changes are provisioned to AD and user can login with new password.
      b .Password reset self service. - User resets password in IDM --> password changes in AD (as UME is configured to use AD)
    Change password (by user)
    By default the users who are successfully authenticated when they try to login to IDM UI, will get access to self-services tab. To allow users to change the password on their own, you create the corresponding ordered tasks and maintain the access control tab for selfservice.
    So that when users wants to change their passwords, they can change on their own.
    How IDM will provision the new password to target system is something you have to configure the logic. For example, my sandbox looks like this.
    Password reset self-service.
    The user can reset their password on their own if they cannot remember their password. To implement this, look at this document. http://scn.sap.com/docs/DOC-17111
    Hope this helps. please let me know for any further queries.
    ~ Krishna.

  • Initial Load is not starting

    Hi,
    On target , in accounts table there are three additional columns namely ,
    S_JOB_ID NUMBER(19)
    S_REPLICATED_DATE DATE
    S_EXTRACT_IND NUMBER(5)
    Out of this , S_JOB_ID should have null value. the value will be inserted by ETL later on. This is our business requirment.
    I tried two options
    COLMAP ( USEDEFAULTS , S_JOB_ID = ' ',
    and
    COLMAP ( USEDEFAULTS , S_JOB_ID = NULL,
    but none of them worked.
    Please help me.
    Thanks,
    Kamlesh

    Well, an empty string is similar to null in some ways, but technically, an empty string is not null, so don't go with that approach.
    Workaround - If you ignore that column and just try to update the other two (i.e., your target table does not have the S_JOB_ID column), does the initial load work? If so, you can alter the target table afterwards, add the column, default value is null.
    Can you post your parameter files?

  • Best Practice for Initial Load

    Hello,
    what is the best way of doing the initial load? is there a best practice somwhere that tells you what should be imported first?
    I want to understand the order ex,
    1. load Lookups,
    2. Hierarchies,
    3. taxonomy and attributes
    last the main table
    etc...
    I dont understand the logic.
    Thanks in advance

    Hi Ario,
    If you follow any SAP Standard business content for MDM Repositories like e.g. Material.
    https://websmp130.sap-ag.de/sap/support/notes/1355137
    In the SAP Note attachments, you will get MDM71_Material_Content.pdf
    You will see Import of reference Data(look up table's data) 1st(step6) before import of Master data(step7).
    During Import of Reference Data(look up data), Please follow the Import Sequence by using Processing level 0,1,2 etc.
    Which take care of filling look up flat tables first then filling Hierarchies tables etc.
    After that if you are maintaining Taxonomy, You need to fill taxonomy table in Taxonomy mode of Data Manager, in the sequence (Categories, Attributes, Linkage between Attributes and Categories and lastly Attribute Values)
    After this I mean populating Reference data you need to populate Main table records along with tuples table data since now in MDM 7.1 Tuple has been replaced by Qualified table for most of the Master's but if you are still maintaining Qualified table you can import Mani table data along with Qualified table in a single step. Otherwise for Qualified table you can alos use this approach of populating Non-qualifeirs to Qualified table first before importing main table and then importing Main table data along with Qualifier's field of Qualified table.
    This above entire process for exporting data from SAP R/3 system to MDM. If you are importing data into MDM from legacy system (Non-Sap systems too), Approach should be remain same Populating Lookup tables data and lastly main table data.
    I dont understand the logic.
    The logic is simple in your main table you have fields which are look up to Reference tables( e.g. field in main table which are look up to Lookup flat tables like Countries, Currencies etc, field in main table which is lookup to Hierarchy/Taxonomy table etc). So, if these values are not populated firstly, so during your Main table import you will have incomplete data for all of these fields from main table which are look up to some other tables as values in your lookup table you haven't populated before Main table import.
    Kindly revert if you still have any doubts.
    Regards,
    Mandeep Saini

  • No initial load of Customers, Material and delta load of Sales Orders.

    Hi Experts,
    I am facing a very troublesome issue. I am not able to setup the Middleware portion for initial and delta loads. I read a lot of documents and corrected a lot of things. finally, the connectivity is done with R/3 and CRM. Initial load of all objects is successful (as per Best PRactices guide). Customizing load is successful.
    But after now I have these open issues for which I am unable to find any answers (am really exhausted!!):
    - Customer_main load, it was succesful, but no BP's of R/3 are available.
    - Material, it failed in SMW01, SMQ2, the errors are:
    Mat. for Initial Download: Function table not supported
    EAN xxxxxxxxxxxxxxxxxx does not correspond to the GTIN format and cannot be transferred
    EAN yyyyyyyyyyyyyyyyyy does not correspond to the GTIN format and cannot be transferred
    Plant xx is not assigned to a business partner
    - Sales order, it shows green bdoc, but error segments says "No upload to R/3" and the order does not flow to R/3.
    We had our system setup 3 years back for data transfer and Middleware. But few things changed and connectivity stopped. I did all that again now, but am not yet successful. Any inputs will be greatly appreciated.
    Thanks,
    -Pat

    Hi Ashvin,
    The error messages in SMW01 for MAterial initial load is :
         Mat. for Initial Download: Function table not supported
         EAN 123456789000562 does not correspond to the GTIN format and cannot be transferred
         EAN 900033056531434 does not correspond to the GTIN format and cannot be transferred
         Plant 21 is not assigned to a business partner
    I have done the DNL_PLANT load successfully. Why then the plant error?
    Some of the messages for BP:
    Messages for business partner 1331:
    No classification is assigned to business partner 1331
    For another,
         Partner 00001872(469206A60E5F61C6E10000009F70045E): the following errors occurred
         City Atlanta does not exist in country US
         Time zone EST_NA does not exist
         You are not allowed to enter a tax jurisdiction code for country US
         Validation error occurred: Module CRM_BUPA_MAIN_VAL, BDoc type BUPA_MAIN.
    Now, the time zone EST is assigned by default in R/3. Where do I change that? I do not want to change time zones as this may have other impacts. Maybe CRM I cna change this, not for sure in R/3. City check has been deactivated in R/3 and CRM, still the error.
    Till these 2 are not solved, I cannot go into the Sales order loads.
    Any ideas will be greatly appreciated.
    Thanks,
    -Pat

  • Initial load of sales orders from R3 to CRM without statuses

    1) Some sales orders were uploaded into CRM without statuses in the headers or line items. 
    2) Some sales orders were uploaded without status, ship-to, sold-to, payer.....If I deleted them and use R3AR2, R3AR4 to upload each individual then no problem.
    Any ideas or suggestions?
    Thanks.

    Hi,
       Request load of adapter objects uses different extractor modules for extracting the data from external system to CRM. While your initial load of sales docs. will use a different extraction logic bases on the filter conditions specfied on trx.
    R3AC1
       There may be a problem in the extraction of data from the source system (don't know if you are using a R/3). Can you please de-register the R/3 (i suppose) outbound queue using trx.
    SMQS
    , and then debug the extraction (R/3 outbound) before the data is sent to CRM using FM
    CRS_SEND_TO_SERVER
       If this goes well, you may try debugging the mapper in CRM inbound and the validation module in CRM as a last resort. Also, please refer to trx.
    SMW01
    to see if the Bdocs are fully processed.
    Hope this helps...Reward if helpful.
    Regards,
    Sudipta.

  • After Upgrade BI Initial load is taking much time

    Dear Friends,
    We had BW 3.5 on  Windows 2003(32 bit) & SQL 2000.
    I upgraded it to BI 7.01 (EHP1 SR1) with Windows 2003(64bit),SQL 2005 & completed all followup activities.
    Now when we are doing initial load it is taking long time. Please do let me know your inputs as soon as possible.
    Regards,
    Sunil Maurya.

    Hi ,
    I created the thread under netweaver forum.
    But still I'll check & try create it in correct forum.
    Regards,
    Sunil Maurya

  • How to use more than one application server during initial load?

    Hi,
    we plan to use more than one application server in CRM during initial download in order to increase the number of parallel requests and to decrease the time for the initial load. Is there a way to allocate requests to more than one server? Is is possible via multiple rfc connections for consumer CRM in CRMRFCPAR?
    Thanks.
    Alexander Schiffer

    Hi Naresh,
    thanks for your answer. It has solved my problem. SMLG is the transaction that I was looking for.
    Two more OSS notes that helped me to guide our basis into the right direction:
    OSS 593058 - New RFC load balancing procedure
    OSS 1413986 - SMLG: Possibility to select a favorite type for Ext.RFCs
    Thanks again.
    Alexander Schiffer

  • Initial load failing between identical tables. DEFGEN skewed and fixable?

    Initial load failing between identical tables. DEFGEN skewed and fixable?
    Error seen:
    2013-01-28 15:23:46 WARNING OGG-00869 [SQL error 0 (0x0)][HP][ODBC/MX Driver] DATETIME FIELD OVERFLOW. Incorrect Format or Data. Row: 1 Column: 11.
    Then compared the discard record against a select * on the key column.
    Mapping problem with insert record (target format)...
    **** Comparing Discard contents to Select * display
    ABCHID = 3431100001357760616974974003012 = 3431100001357760616974974003012
    *!!! ABCHSTEPCD = 909129785 <> 9 ???*
    ABCHCREATEDDATE = 2013-01-09 13:43:36 = 2013-01-09 13:43:36
    ABCHMODIFIEDDATE = 2013-01-09 13:43:36 =2013-01-09 13:43:36
    ABCHNRTPUSHED = 0 = 0
    ABCHPRISMRESULTISEVALUATED = 0 = 0
    SABCHPSEUDOTERM = 005340 = 005340
    ABCHTERMID = TERM05 = TERM05
    ABCHTXNSEQNUM = 300911112224 = 300911112224
    ABCHTIMERQSTRECVFROMACQR = 1357799914310 = 1357799914310
    *!!! ABCTHDATE = 1357-61-24 00:43:34 <> 2013-01-09 13:43:34*
    ABCHABCDATETIME = 2013-01-09 13:43:34.310000 = 2013-01-09 13:43:34.310000
    ABCHACCOUNTABCBER =123ABC = 123ABC
    ABCHMESSAGETYPECODE = 1210 = 1210
    ABCHPROCCDETRANTYPE = 00 = 00
    ABCHPROCCDEFROMACCT = 00 = 00
    ABCHPROCCDETOACCT = 00 = 00
    ABCHRESPONSECODE = 00 = 00
    …. <snipped>
    Defgen comes out same when run against either table.
    Also have copied over and tried both outputs from DEFGEN.
    +- Defgen version 2.0, Encoding ISO-8859-1
    * Definitions created/modified 2013-01-28 15:00
    * Field descriptions for each column entry:
    * 1 Name
    * 2 Data Type
    * 3 External Length
    * 4 Fetch Offset
    * 5 Scale
    * 6 Level
    * 7 Null
    * 8 Bump if Odd
    * 9 Internal Length
    * 10 Binary Length
    * 11 Table Length
    * 12 Most Significant DT
    * 13 Least Significant DT
    * 14 High Precision
    * 15 Low Precision
    * 16 Elementary Item
    * 17 Occurs
    * 18 Key Column
    * 19 Sub Data Type
    Database type: SQLMX
    Character set ID: ISO-8859-1
    National character set ID: UTF-16
    Locale: en_EN_US
    Case sensitivity: 14 14 14 14 14 14 14 14 14 14 14 14 11 14 14 14
    Definition for table RT.ABC
    Record length: 1311
    Syskey: 0
    Columns: 106
    ABCHID 64 34 0 0 0 0 0 34 34 34 0 0 32 32 1 0 1 3
    ABCHSTEPCD 132 4 39 0 0 0 0 4 4 4 0 0 0 0 1 0 0 0
    ABCHCREATEDDATE 192 19 46 0 0 0 0 19 19 19 0 5 0 0 1 0 0 0
    ABCHMODIFIEDDATE 192 19 68 0 0 0 0 19 19 19 0 5 0 0 1 0 0 0
    ABCHNRTPUSHED 130 2 90 0 0 0 0 2 2 2 0 0 0 0 1 0 0 0
    ABCHPRISMRESULTISEVALUATED 130 2 95 0 0 0 0 2 2 2 0 0 0 0 1 0 0 0
    ABCHPSEUDOTERM 0 8 100 0 0 0 0 8 8 8 0 0 0 0 1 0 0 0
    ABCTERMID 0 16 111 0 0 0 0 16 16 16 0 0 0 0 1 0 0 0
    ABCHTXNSEQNUM 0 12 130 0 0 0 0 12 12 12 0 0 0 0 1 0 0 0
    ABCHTIMERQSTRECVFROMACQR 64 24 145 0 0 0 0 24 24 24 0 0 22 22 1 0 0 3
    ABCTHDATE 192 19 174 0 0 0 0 19 19 19 0 5 0 0 1 0 0 0
    ABCHABCDATETIME 192 26 196 0 0 1 0 26 26 26 0 6 0 0 1 0 0 0
    ABCHACCOUNTABCER 0 19 225 0 0 1 0 19 19 19 0 0 0 0 1 0 0 0
    ABCHMESSAGETYPECODE 0 4 247 0 0 1 0 4 4 4 0 0 0 0 1 0 0 0
    ABCHPROCCDETRANTYPE 0 2 254 0 0 1 0 2 2 2 0 0 0 0 1 0 0 0
    ABCHPROCCDEFROMACCT 0 2 259 0 0 1 0 2 2 2 0 0 0 0 1 0 0 0
    ABCHPROCCDETOACCT 0 2 264 0 0 1 0 2 2 2 0 0 0 0 1 0 0 0
    ABCHRESPONSECODE 0 5 269 0 0 1 0 5 5 5 0 0 0 0 1 0 0 0
    … <snipped>
    The physical table shows a PACKED REC 1078
    And table invoke is:
    -- Definition of table ABC3.RT.ABC
    -- Definition current Mon Jan 28 18:20:02 2013
    ABCHID NUMERIC(32, 0) NO DEFAULT HEADING '' NOT
    NULL NOT DROPPABLE
    , ABCHSTEPCD INT NO DEFAULT HEADING '' NOT NULL NOT
    DROPPABLE
    , ABCHCREATEDDATE TIMESTAMP(0) NO DEFAULT HEADING '' NOT
    NULL NOT DROPPABLE
    , ABCHMODIFIEDDATE TIMESTAMP(0) NO DEFAULT HEADING '' NOT
    NULL NOT DROPPABLE
    , ABCHNRTPUSHED SMALLINT DEFAULT 0 HEADING '' NOT NULL NOT
    DROPPABLE
    , ABCHPRISMRESULTISEVALUATED SMALLINT DEFAULT 0 HEADING '' NOT NULL NOT
    DROPPABLE
    , ABCHPSEUDOTERM CHAR(8) CHARACTER SET ISO88591 COLLATE
    DEFAULT NO DEFAULT HEADING '' NOT NULL NOT DROPPABLE
    , ABCHTERMID CHAR(16) CHARACTER SET ISO88591 COLLATE
    DEFAULT NO DEFAULT HEADING '' NOT NULL NOT DROPPABLE
    , ABCHTXNSEQNUM CHAR(12) CHARACTER SET ISO88591 COLLATE
    DEFAULT NO DEFAULT HEADING '' NOT NULL NOT DROPPABLE
    , ABCHTIMERQSTRECVFROMACQR NUMERIC(22, 0) NO DEFAULT HEADING '' NOT
    NULL NOT DROPPABLE
    , ABCTHDATE TIMESTAMP(0) NO DEFAULT HEADING '' NOT
    NULL NOT DROPPABLE
    , ABCHABCDATETIME TIMESTAMP(6) DEFAULT NULL HEADING ''
    , ABCHACCOUNTNABCBER CHAR(19) CHARACTER SET ISO88591 COLLATE
    DEFAULT DEFAULT NULL HEADING ''
    , ABCHMESSAGETYPECODE CHAR(4) CHARACTER SET ISO88591 COLLATE
    DEFAULT DEFAULT NULL HEADING ''
    , ABCHPROCCDETRANTYPE CHAR(2) CHARACTER SET ISO88591 COLLATE
    DEFAULT DEFAULT NULL HEADING ''
    , ABCHPROCCDEFROMACCT CHAR(2) CHARACTER SET ISO88591 COLLATE
    DEFAULT DEFAULT NULL HEADING ''
    , ABCHPROCCDETOACCT CHAR(2) CHARACTER SET ISO88591 COLLATE
    DEFAULT DEFAULT NULL HEADING ''
    , ABCHRESPONSECODE CHAR(5) CHARACTER SET ISO88591 COLLATE
    DEFAULT DEFAULT NULL HEADING ''
    …. Snipped
    I suspect that the fields having subtype 3 just before the garbled columns is a clue, but not sure what to replace with or adjust.
    Any and all help mighty appreciated.

    Worthwhile suggestion, just having difficulty applying.
    I will tinker with it more. But still open to more suggestions.
    =-=-=-=-
    Oracle GoldenGate Delivery for SQL/MX
    Version 11.2.1.0.1 14305084
    NonStop H06 on Jul 11 2012 14:11:30
    Copyright (C) 1995, 2012, Oracle and/or its affiliates. All rights reserved.
    Starting at 2013-01-31 15:19:35
    Operating System Version:
    NONSTOP_KERNEL
    Version 12, Release J06
    Node: abc3
    Machine: NSE-AB
    Process id: 67895711
    Description:
    ** Running with the following parameters **
    2013-01-31 15:19:40 INFO OGG-03035 Operating system character set identified as ISO-8859-1. Locale: en_US_POSIX, LC_ALL:.
    Comment
    Comment
    REPLICAT lodrepx
    ASSUMETARGETDEFS
    Source Context :
    SourceModule : [er.init]
    SourceID : [home/ecloud/sqlmx_mlr14305084/src/app/er/init.cpp]
    SourceFunction : [get_infile_params]
    SourceLine : [2418]
    2013-01-31 15:19:40 ERROR OGG-00184 ASSUMETARGETDEFS is not supported for SQL/MX ODBC replicat.
    2013-01-31 15:19:45 ERROR OGG-01668 PROCESS ABENDING.

  • Error while Initial load

    Hi Forum,
    Iam doing Middleware setup for downloading customer master from R/3 to CRM.Iam trying to do initial load of customizing objects viz.,DNL_CUST_ACGRPB,DNL_CUST_ADDR,
    DNL_CUST_KTOKD,DNL_CUST_TVKN,DNL_CUST_TVLS,DNL_CUST_TVPV... which are to be loaded before doing initial load of CUSTOMER_MAIN.
    while doing initial load of customizing objects iam getting the below mentioned error
    <b>001 No generation performed. Call transaction GN_START.</b>
    <b>002 Due to system errors the Load is prohibited (check transaction MW_CHECK)!</b>
    <b>-</b>when I do GN_START
    "A  job is already scheduled periodically.
    Clicking on 'Continue' will create another job
    that starts immediately.
    Do you want to continue?" message is displayed
    and I have sheduled it.
    But in SMWP transaction I can see in
    <b>BDoc Types: Generation of other runtime objects</b>
    Not generated / <b>generated with errors 2 entries  
                    31.08.2006 05:33:50</b>
    and the objects with errors are
    <b>POT_LISTWRITE
    SPE_DDIC_WRITE</b>
    <b>-</b>In transaction MW_CHECK, system displays message as <b>No generation performed. Call transaction GN_START.</b>     
    when I regenerate these objects(generated with errors)from the context menu I find no difference.
    I have also referred to the <b>Note :637836 and 661067</b> which also suggests to run few reports and GN_START but inspite of doing all the corretion parameters in the note Iam still unable to come out of the situation.
    Please Guide
    Thanks in Advance
    Shridhar.
    Message was edited by: Shridhar Deshpande

    Hi Rahul,
    Thanks for the reply.I checked in transaction MW_CHECK and the system throws the message as
    <b>No generation performed. Call transaction GN_START.</b>
    In the long text the below message is available
    <b>No generation performed. Call transaction GN_START.
    Message no. SMW_GEN_KERNEL005
    Diagnosis
    An upgrade was performed.
    <b>System response</b>
    The Middleware is stopped because MW objects must be generated.
    <b>Procedure</b>
    Excecute transaction GN_START.</b>
    If GN_START is executed,I dont find any change.
    I also checked in <b>smq2</b>in CRM and I found the status of the queue as below
        CL       Queue Name      Entries  Status   Date
    <b>200 CSABUPA0000000042  5       SYSFAIL  31.08.2006 09:54:05 31.08.2006 09:54:11</b>
    Thanks
    Shridhar
    Message was edited by: Shridhar Deshpande

  • Connection Refused Error while running AS ABAP Initial Load

    All,
    I've never connected SAP NW IdM to an actual SAP system before, and I feel like I'm missing some obvious step of configuration, but I can't figure out what.
    We are in the process of trying provision user accounts to our SAP ABAP systems. My first step was to try to read all of the existing accounts from the ABAP system:
    Our Basis team created me a Communication user with the proper authorizations (I ensured that the authorizations included in SAP_BC_SEC_IDM_COMMUNICATION)
    I created a repository using the SAP NewWeaver AS ABAP (Specific Application Server) Repository Template (No CUA, No SNC) using that user's credentials
    I then used the Job Wizard and used the job template AS ABAP - Initial Load, specifying my repository above.
    When I run the job I get the following:
    Initializing SAP connection with parameters:
    com.sap.mw.jco.JCO$Exception: (102) RFC_ERROR_COMMUNICATION: Connect to message server failed Connect_PM MSHOST=<IPADDRESS>, R3NAME=<SID>, GROUP=PUBLIC LOCATION CPIC (TCP/IP) on local host ERROR partner not reached (host <IPADDRESS>, service sapms<SID>) TIME Wed Jun 01 13:54:36 201 RELEASE 640 COMPONENT NI (network interface) VERSION 37 RC -10 MODULE nixxi_r.cpp LINE 8840 DETAIL NiPConnect2 SYSTEM CALL SiPeekPendConn ERRNO 10061 ERRNO TEXT WSAECONNREFUSED: Connection refused COUNTER 1
    Error Init failed
    I'm running SAP NW IdM 7.1 SP5 on Windows Server 2003 with MSSQL 2005. The ABAP server is on a UNIX box with an Oracle 10.2 dB.
    Is there additional configuration that needs to take place on the ABAP side to accept the connection?
    I've tried to find documentation on this, but have been unsuccessful. If someone could point me to the correct documents, or at least point me in the right direction for troubleshooting, it would be greatly appreciated.

    Ankur,
    Looks like the endpoint URL of the webservice is not updated and still pointing to the localhost. Try changing it to http://<ip_addres_of_your_server>:7101..... and see if it works fine then.
    -Arun

  • Sales document types during initial load

    Hi Gurus,
    Landscape - SAP R/3 4.6 c, CRM - 4.0
    We are starting the initial load for the Sales documents, item categories etc.
    For the same, for the document types in R/3, we have the same document type in CRM.
    Question 1- Do I need to configure the partner determination, incompletion log etc for the document type before the initial load or I can do it later once the data is in CRM.
    Question 2- Is it mandatory that for all the doc types in R/3, we have the respective document type in CRM? Logically yes, if yes, where can I find the mapping? For example for OR in R/3, there is a TA in CRM.
    Request someone to answer this ASAP.
    thanks
    Yash

    Hi Yash,
    I think the behavior is just use the same code
    of Transaction Type in CRM and Order Type in R/3.
    Since partner determiniation using different partner
    function codes in CRM and R/3, maintained or not in R/3,
    you may still need to maintain it again in CRM.
    For document, what you can do just create same document
    type (Transaction type) for R/3 and CRM. And depends upon
    you sceneario (in CRM there are several scenarios such as
    document created in CRM and transfered to R/3 Or download
    R/3 document into CRM - no transaction in CRM).
    Especially for partner, you may need to setup data
    transfer CRM to R/3 or R/3 to CRM in CRM configuration.
    And in R/3 you need to define it in tcode PIDE.
    Hope this helps.
    Cheers,
    Gun.

  • Replicating data once again to CRM after initial load fails for few records

    My question (to put it simply):
    We performed an initial load for customers and some records error out in CRM due to invalid data in R/3. How do we get the data into CRM after fixing the errors in R/3?
    Detailed information:
    This is a follow up question to the one posted here.
    Can we turn off email validation during BP replication ?
    We are doing an initial load of customers from R/3 to CRM, and those customers with invalid email address in R/3 error out and show up in SMW01 as having an invalid email address.
    If we decide to fix the email address errors on R/3, these customers should then be replicated to CRM automatically, right? (since the deltas for customers are already active) The delta replication takes place, but, then we get this error message "Business Partner with GUID 'XXXX...' does not exist".
    We ran the program ZREPAIR_CRMKUNNR provided by SAP to clear out any inconsistent data in the intermediate tables CRMKUNNR and CRM_BUT_CUSTNO, and then tried the delta load again. It still didn't seem to go through.
    Any ideas how to resolve this issue?
    Thanks in advance.
    Max

    Subramaniyan/Frederic,
    We already performed an initial load of customers from R/3 to CRM. We had 30,330 records in R/3 and 30,300 of them have come over to CRM in the initial load. The remaining 30 show BDOC errors due to invalid email address.
    I checked the delta load (R3AC4) and it is active for customers. Any changes I make for customers already in CRM come through successfully.  When I make changes to customers with an invalid email address, the delta gets triggered and data come through to CRM, and I get the BDOC error "BP with GUID XXX... does not exist"
    When I do a request load for that specific customer, it stays in "Wait" state forever in "Monitor Requests"
    No, the DIMA did not help Frederic. I did follow the same steps you had mentioned in the other thread, but it just doesn't seem to run. I am going to open an OSS message with SAP for it. I'll update the other thread.
    Thanks,
    Max

Maybe you are looking for