Payroll Conversions to load tables T558B,T558C, T5U8C -- Please Suggest

Hi All
We are currently trying to do data conversion at a client with 100,000 + Employees. We are trying load the tables
T558B - Payroll Periods
T558C - Payroll Account Transfer: Old Wage Types
T5U8C - Transfer external payroll results (USA)
for all the YTD taxes for the Employees.
We have designed LSMW's using Transaction code SM30 to load all the Employees. When we are loading using SM30, it is taking a very long time. Can anyone suggest, if this is the Best practice, or if there is an alternate way to load all these data into SAP without using SM30.
Please let me know
Thank you
Deepthi
Note: I have checked SAP Best practice and that document uses SM30 transaction recording

Hi
We are facing the same problem can you please tell us how did you resolve
Venkat

Similar Messages

  • Error while impdp: ORA-02374: conversion error loading table

    Hi,
    I am trying to convert the character set from WE8ISO8859P1 to AL32UTF8 using expdp/impdp. for this I first convert WE8ISO8859P1 to WE8MSWIN1252 in source DB to get rid of “lossy” data. I created new database(target) with character set AL32UTF8 and nls_length_semantics = ’CHAR’ and created all the tablespaces as in source DB with auoexend on. I took full export (expdp) of source DB excluding TABLESPACE,STATISTICS,INDEX,CONSTRAINT,REF_CONSTRAINT and imported using impdp to target DB. I found below error in the import log file:
    ORA-02374: conversion error loading table "SCTCVT"."SPRADDR_CVT"
    ORA-26093: input data column size (44) exceeds the maximum input size (40)
    ORA-02372: data for row: CONVERT_STREET_LINE1 : 0X'20202020202020202020202020202020202020202020202020'
    I checked with select query on both DBs with below results.
    source DB:
    04:58:42 SQL> select count(*) from "SCTCVT"."SPRADDR_CVT";
    COUNT(*)
    74553
    target DB:
    04:59:24 SQL> select count(*) from "SCTCVT"."SPRADDR_CVT";
    COUNT(*)
    74552
    please suggest me a solution to this.
    Thanks and Regards.
    Edited by: user12045167 on May 9, 2011 10:39 PM

    Thanks for your update maher.
    09:15:53 SQL> desc "SCTCVT"."SPRADDR_CVT"
    Name Null? Type
    SPRADDR_PIDM NUMBER(8)
    CONVERT_PIDM VARCHAR2(9 CHAR)
    SPRADDR_ATYP_CODE VARCHAR2(2 CHAR)
    CONVERT_ATYP_CODE VARCHAR2(2 CHAR)
    SPRADDR_SEQNO NUMBER(2)
    CONVERT_SEQNO VARCHAR2(2 CHAR)
    SPRADDR_FROM_DATE DATE
    CONVERT_FROM_DATE VARCHAR2(8 CHAR)
    SPRADDR_TO_DATE DATE
    CONVERT_TO_DATE VARCHAR2(8 CHAR)
    SPRADDR_STREET_LINE1 VARCHAR2(30 CHAR)
    CONVERT_STREET_LINE1 VARCHAR2(40 CHAR)
    SPRADDR_STREET_LINE2 VARCHAR2(30 CHAR)
    CONVERT_STREET_LINE2 VARCHAR2(40 CHAR)
    SPRADDR_STREET_LINE3 VARCHAR2(30 CHAR)
    CONVERT_STREET_LINE3 VARCHAR2(40 CHAR)
    SPRADDR_CITY VARCHAR2(20 CHAR)
    CONVERT_CITY VARCHAR2(25 CHAR)
    SPRADDR_STAT_CODE VARCHAR2(3 CHAR)
    CONVERT_STAT_CODE VARCHAR2(25 CHAR)
    SPRADDR_ZIP VARCHAR2(10 CHAR)
    CONVERT_ZIP VARCHAR2(15 CHAR)
    SPRADDR_CNTY_CODE VARCHAR2(5 CHAR)
    CONVERT_CNTY_CODE VARCHAR2(5 CHAR)
    SPRADDR_NATN_CODE VARCHAR2(5 CHAR)
    CONVERT_NATN_CODE VARCHAR2(5 CHAR)
    SPRADDR_PHONE_AREA VARCHAR2(3 CHAR)
    CONVERT_PHONE_AREA VARCHAR2(3 CHAR)
    SPRADDR_PHONE_NUMBER VARCHAR2(7 CHAR)
    CONVERT_PHONE_NUMBER VARCHAR2(7 CHAR)
    SPRADDR_PHONE_EXT VARCHAR2(4 CHAR)
    CONVERT_PHONE_EXT VARCHAR2(4 CHAR)
    SPRADDR_STATUS_IND VARCHAR2(1 CHAR)
    CONVERT_STATUS_IND VARCHAR2(1 CHAR)
    SPRADDR_ACTIVITY_DATE DATE
    CONVERT_ACTIVITY_DATE VARCHAR2(8 CHAR)
    SPRADDR_USER VARCHAR2(30 CHAR)
    CONVERT_USER VARCHAR2(30 CHAR)
    SPRADDR_ASRC_CODE VARCHAR2(4 CHAR)
    CONVERT_ASRC_CODE VARCHAR2(4 CHAR)
    SPRADDR_DELIVERY_POINT NUMBER(2)
    CONVERT_DELIVERY_POINT VARCHAR2(2 CHAR)
    SPRADDR_CORRECTION_DIGIT NUMBER(1)
    CONVERT_CORRECTION_DIGIT VARCHAR2(1 CHAR)
    SPRADDR_CARRIER_ROUTE VARCHAR2(4 CHAR)
    CONVERT_CARRIER_ROUTE VARCHAR2(4 CHAR)
    SPRADDR_GST_TAX_ID VARCHAR2(15 CHAR)
    CONVERT_GST_TAX_ID VARCHAR2(15 CHAR)
    SPRADDR_REVIEWED_IND VARCHAR2(1 CHAR)
    CONVERT_REVIEWED_IND VARCHAR2(1 CHAR)
    SPRADDR_REVIEWED_USER VARCHAR2(30 CHAR)
    CONVERT_REVIEWED_USER VARCHAR2(30 CHAR)
    SPRADDR_DATA_ORIGIN VARCHAR2(30 CHAR)
    CONVERT_DATA_ORIGIN VARCHAR2(30 CHAR)
    SPRADDR_CVT_RECORD_ID NUMBER(8)
    SPRADDR_CVT_STATUS VARCHAR2(1 CHAR)
    SPRADDR_CVT_JOB_ID NUMBER(8)
    so here we can see its value is 40 (CONVERT_STREET_LINE1 VARCHAR2(40 CHAR)).
    shall i go ahead altering the column?

  • ORA-02374: conversion error loading table during import using IMPDP

    HI All,
    We are trying to migrate the data from one database to an other database.
    The source database is having character set
    SQL> select value from nls_database_parameters where parameter='NLS_CHARACTERSET';
    VALUE
    US7ASCII
    The destination database is having character set
    SQL> select value from nls_database_parameters where parameter='NLS_CHARACTERSET';
    VALUE
    AL32UTF8
    We took an export of the whole database using expdp and when we try to import to the destination database using impdp. We are getting the following error.
    ORA-02374: conversion error loading table <TABLE_NAME>
    ORA-12899: value too large for column <COLUMN NAME> (actual: 42, maximum: 40)
    ORA-02372: data for row:<COLUMN NAME> : 0X'4944454E5449464943414349E44E204445204C4C414D414441'
    Kindly let me know how to overcome this issue in destination.
    Thanks & Regards,
    Vikas Krishna

    Hi,
    You can overcome this issue by increasing the column width in the target database for the max value required for all data to be imported successfully in the table.
    Regards

  • I had run a  query  I want to insert that query result into another table how it is please suggest me

    1
    1
    EC101
    P
    15
    P
    40
    2010-11-01
    1
    2
    1
    EC102
    P
    10
    P
    30
    2010-11-01
    1
    3
    1
    EC103
    P
    8
    P
    50
    2010-11-01
    1
    4
    1
    EC104
    P
    5
    P
    25
    2010-11-01
    1
    5
    1
    EC105
    P
    10
    P
    40
    2010-11-01
    1
    6
    1
    EC106
    P
    12
    P
    45
    2010-11-01
    1
    7
    2
    EC101
    P
    20
    P
    50
    2010-11-01
    1
    8
    2
    EC102
    P
    15
    P
    60
    2010-11-01
    1
    9
    2
    EC103
    P
    20
    P
    80
    2010-11-01
    1
    10
    2
    EC104
    P
    14
    P
    46
    2010-11-01
    1
    11
    2
    EC105
    P
    18
    P
    74
    2010-11-01
    1
    12
    2
    EC106
    P
    16
    P
    68
    2010-11-01
    1
    13
    3
    EC101
    P
    10
    P
    40
    2010-11-01
    1
    14
    3
    EC102
    P
    20
    P
    60
    2010-11-01
    1
    15
    3
    EC103
    P
    12
    P
    40
    2010-11-01
    1
    16
    3
    EC104
    P
    20
    P
    50
    2010-11-01
    1
    17
    3
    EC105
    P
    12
    P
    40
    2010-11-01
    1
    18
    3
    EC106
    P
    14
    P
    60
    2010-11-01
    1
    19
    4
    EC101
    P
    10
    P
    30
    2010-11-01
    1
    20
    4
    EC102
    P
    14
    P
    40
    2010-11-01
    1
    21
    4
    EC103
    P
    12
    P
    50
    2010-11-01
    1
    22
    4
    EC104
    P
    8
    P
    46
    2010-11-01
    1
    23
    4
    EC105
    P
    4
    P
    25
    2010-11-01
    1
    24
    4
    EC106
    P
    16
    P
    68
    2010-11-01
    1
    25
    1
    EC201
    P
    10
    P
    42
    2011-04-01
    2
    26
    1
    EC202
    P
    18
    P
    60
    2011-04-01
    2
    27
    1
    EC203
    P
    8
    P
    50
    2011-04-01
    2
    28
    1
    EC204
    P
    10
    P
    40
    2011-04-01
    2
    29
    1
    EC205
    P
    8
    P
    30
    2011-04-01
    2
    30
    1
    EC206
    P
    20
    P
    50
    2011-04-01
    2
    31
    2
    EC201
    P
    10
    P
    40
    2011-04-01
    2
    32
    2
    EC202
    P
    12
    P
    45
    2011-04-01
    2
    33
    2
    EC203
    P
    12
    P
    46
    2011-04-01
    2
    34
    2
    EC204
    P
    20
    P
    40
    2011-04-01
    2
    35
    2
    EC205
    P
    10
    P
    40
    2011-04-01
    2
    36
    2
    EC206
    P
    10
    P
    48
    2011-04-01
    2
    37
    3
    EC201
    P
    20
    P
    60
    2011-04-01
    2
    38
    3
    EC202
    P
    18
    P
    70
    2011-04-01
    2
    39
    3
    EC203
    P
    15
    P
    80
    2011-04-01
    2
    40
    3
    EC204
    P
    18
    P
    78
    2011-04-01
    2
    41
    3
    EC205
    P
    16
    P
    65
    2011-04-01
    2
    42
    3
    EC206
    P
    10
    P
    30
    2011-04-01
    2
    43
    4
    EC201
    P
    20
    P
    50
    2011-04-01
    2
    44
    4
    EC202
    P
    14
    P
    10
    2011-04-01
    2
    45
    4
    EC203
    P
    15
    P
    40
    2011-04-01
    2
    46
    4
    EC204
    P
    20
    P
    46
    2011-04-01
    2
    47
    4
    EC205
    P
    10
    P
    50
    2011-04-01
    2
    48
    4
    EC206
    P
    8
    P
    40
    2011-04-01
    2
    from this table I want take the semester wise result and after  that I want to take that result into new table  and I want take the rank also
    how

    Hey,
    Just perform an INSERT AS SELECT and don't forget to add APPEND hint. Also, you can do it using parallelism so before that just enable DML parallel for the IAS. Below are the commands:
    SQL> ALTER SESSION ENABLE PARALLEL DML;
    SQL> INSERT /*+ APPEND */ INTO NEW_TABLE (SELECT * FROM TABLE_QUERY);
    Regards.
    caadecarvalho

  • T558B, T558C, T558D, T558E, T5K8C, T5K8D, T5K8E, T5U8C, T5U8A

    Hi guys,
    In need to load contents of tables (T558B, T558C, T558D, T558E, T5K8C, T5K8D, T5K8E, T5U8C, T5U8A) . Could you please tell me which method is the SAP recommended way for the loading?
    Thanks a lot,
    Franck

    Hi man,
    The most common method is through LSMW (T-Code). There you can upload information into structures/tables.
    You will find more detailed information using the following link:
    Use of Table T556C
    The market offers some other tools that also help with the load and are friendly to use. It depends on your budget and timing.
    Regards.

  • Payroll Conversions

    Hey can anyone help me in Payroll Conversions.
    All i know about this is loading T tables from legacy data.
    T558B
    T558C
    T5u8C....please help me on this...
    Thankx in advance..

    Hi
    We are facing the same problem can you please tell us how did you resolve
    Venkat

  • T558B & T558C

    Hi everyone
    I need to know, what do I have to do, after uplod tables T558B & T558C to make a history cluster,  before a golive in payroll.
    Please, I am working in molga 37, if you have any experience, or if I have to run some international schema, and what do I have to take care, like PA03.
    Regards,
    Rubé

    hi,
    after you get the history in those tables, you need to run the payroll driver program with schema ULK9 which stores the payroll results history including tax results to the RT and TCRT table.
    may be this will help you.
    regards,
    sam..

  • Conversion error in Table control for Negative values

    hello all,
    Iam using table control with 6 columns and in that columns values are coming from Ztable. and out of that one column is of value which is input/output field and having negative value's also.type of that column is DEC.
    Now when i run the report dump occurs showing error message as COnversion Error.
    I think its because of negative value even the in attributes of that column i had selected With Sign checkbox.
    Now when i run the report then if in Ztable negative value is there then i want to display negative value in table control also.
    Please suggest me...
    <removed_by_moderator>
    Thanking you.
    Regards.
    Edited by: Hemant Baviskar on Sep 10, 2008 3:07 PM
    Edited by: Julius Bussche on Sep 10, 2008 1:13 PM

    Hemant
    Refer to following link:
    Table Control : Conversion error
    Thanks
    Amol Lohade

  • Best Practices for Payroll Conversions

    Hi All
    We are currently trying to do data conversion at a client with 100,000 + Employees. We are trying load the tables
    T558B - Payroll Periods
    T558C - Payroll Account Transfer: Old Wage Types
    T5U8C - Transfer external payroll results (USA)
    for all the YTD taxes for the Employees.
    We have designed LSMW's using Transaction code SM30 to load all the Employees. When we are loading using SM30, it is taking a very long time. Can anyone suggest, if this is the Best practice, or if there is an alternate way to load all these data into SAP without using SM30.
    Please let me know
    Thank you
    Deepthi
    Note: I have checked SAP Best practice and that document uses SM30 transaction recording

    Hi,
    In implementing any LO module, generally the following are the points to be taken in mind.
    1. Base level have a ODS to load data from R/3. This ensure that you have exact data content as that in R/3. This can be a write optimized.
    2. Second level have an ODS along with the transformation and modification of data based on the business and functional requirement and enhancments.
    3. Finally have a cube to consolidate the data and make it available for reporting.Create all the reports based on the cube.
    Hope this gives an idea.
    Regards,
    akhan
    Edited by: Akhan_BI on Sep 5, 2009 12:41 AM

  • OBIEE execute stored procedure to load tables before running report

    Hi..
    I want to execute a stored procedure to load database tables before running a report in OBIEE .
    I need to pass 2 parameters to the stored procedure which loads into tables.
    In the Connection Pool --> Connection Script Tab --> Execute before query, I wrote the below query using the repository variables VAR1 & VAR2 to execute the
    DECLARE VAR1 number; VAR2 number;
    BEGIN
    schema_name.package_name1.package_body('VALUE OF(VAR1)', 'VALUE OF(VAR2)'); COMMIT;
    schema_name.package_name2.package_body('VALUE OF(VAR1)', 'VALUE OF(VAR2)'); COMMIT;
    END;
    I am receiving the following error to declare the schema_name.package_name
    +++Administrator:2a0000:2a0004:----2010/06/21 14:29:00
    -------------------- Sending query to database named ACBS-OCC (id: <<49419>>):
    BEGIN schema_name.package_name1.package_body1('VALUE OF(VAR1', 'VALUE OF(VAR2'); COMMIT; schema_name.package_name2.package_body2('VALUE OF(VAR1)', 'VALUE OF(VAR2)'); COMMIT;END;
    +++Administrator:2a0000:2a0004:----2010/06/21 14:29:00
    -------------------- Query Status: Query Failed: [nQSError: 16001] ODBC error state: S1000 code: 6550 message: [Oracle][ODBC][Ora]ORA-06550: line 1, column 7:
    PLS-00201: identifier 'SCHEMA_NAME.PACKAGE_NAME1' must be declared
    ORA-06550: line 1, column 7:
    PL/SQL: Statement ignored
    ORA-06550: line 1, column 93:
    PLS-00201: identifier 'SCHEMA_NAME.PACKAGE_NAME2' must be declared
    ORA-06550: line 1, column 93:
    PL/SQL: Statement ignored.
    [nQSError: 16015] SQL statement execution failed.
    Please suggest how to declare and execute the stored procedure.
    Thanks in advance.

    Hi,
    I know that any Function / Procedure needs to be called using a EVALUATE function in OBIEE.
    Thanks,
    Vijay

  • Error while loading table from flat file (.csv)

    I have a flat file which i am loading into a Target Table in Oracle Warehouse Builder. It uses SQL Loader Internally to load the data from flat file, I am facing an issue. Please find the following error ( This is an extract from the error log generated)
    SQL*Loader-500: Unable to open file (D:\MY CURRENT PROJECTS\GEIP-IHSS-Santa Clara\CDI-OWB\Source_Systems\Acquisition.csv)
    SQL*Loader-552: insufficient privilege to open file
    SQL*Loader-509: System error: The data is invalid.
    SQL*Loader-2026: the load was aborted because SQL Loader cannot continue.
    I believe that this is related to SQL * Loader error.
    ACtually the flat file resides in my system ( D:\MY CURRENT PROJECTS\GEIP-IHSS-Santa Clara\CDI-OWB\Source_Systems\Acquisition.csv). I am connecting to a oracle server.
    Please suggest
    Is it required that i need to place the flat file in Oracle Server System ??
    Regards,
    Ashoka BL

    Hi
    I am getting an error as well which is similar to that described above except that I get
    SQL*Loader-500: Unable to open file (/u21/oracle/owb_staging/WHITEST/source_depot/Durham_Inventory_Labels.csv)
    SQL*Loader-553: file not found
    SQL*Loader-509: System error: The system cannot find the file specified.
    SQL*Loader-2026: the load was aborted because SQL Loader cannot continue.
    The difference is that Ashoka was getting
    SQL*Loader-552: insufficient privilege to open file
    and I get
    SQL*Loader-553: file not found
    The initial thought is that the file does not exist in the directory specified or I have spelt the filename incorrectly it but this has been checked and double checked. The unix directory also has permission to read and write.
    Also in the error message is
    Control File: C:\u21\oracle\owb_staging\WHITEST\source_depot\INV_LOAD_LABEL_INVENTORY.ctl
    Character Set WE8MSWIN1252 specified for all input.
    Data File: /u21/oracle/owb_staging/WHITEST/source_depot/Durham_Inventory_Labels.csv
    Bad File: C:\u21\oracle\owb_staging\WHITEST\source_depot\Durham_Inventory_Labels.bad
    As can be seen from the above it seems to be trying to create the ctl and bad file on my c drive instead of on the server in the same directory as the .csv file. The location is registered to the server directory /u21/oracle/owb_staging/WHITEST/source_depot
    I am at a lost as this works fine in development and I have just promoted all the development work to a systest environment using OMBPlus.
    The directory structure in development is the same as systest except that the data file is /u21/oracle/owb_staging/WHITED/source_depot/Durham_Inventory_Labels.csv and everything works fine - .ctl and .bad created in the same directory and the data sucessfully loads into a oracle table.
    Have I missed a setting in OWB during the promotion to systest or is there something wrong in the way the repository in the systest database is setup?
    The systest and development databases are on the same box.
    Any help would be much appreciated
    Thanks
    Edwin

  • How to do conversion of a table

    while i am removing a primary key of a field of a  Z table , i was asked to do 'conversion of table'. please suggest me, how to do it?

    Hello,
    welcome to the SDN Forums. Please search before you post. A solution for your problem can be found in Topic:
    Custom DB table conversion
    Best regards
    Gregor

  • How to Load table from client file in C?

    Hi all,
    I'm trying to determine how to write a C program to load a client file with the "load table ... using client file..." syntax without calling out to the dbisql program to load the table.  I'm probably not seeing the forest for the trees here.  I know I can use the bulk api (equivalent of bcp) but would rather use the load table because it is much faster.
    jason

    There is nothing special to do - you execute the statement from your program just like any other SQL statement.  The only thing to be aware of are the privilege/permission issues:
    When loading from a file on a client computer:
    READ CLIENT FILE privilege is also required for the database user.
    Read privileges are required on the directory being read from.
    The allow_read_client_file database option must be enabled.
    The read_client_file secure feature must be enabled.
    Revoking these privileges is also the only way you can prevent a user from executing the statement.

  • Error during Unicode Conversion at DDNTF table

    Hello
    We are doing a Unicode Conversion for our CRT systems. This system now
    is a SAP CRM 7.0 SR1 ABAP.
    We have performed preparatory steps (SPUMG, checks, updating R3load,
    R3ldctl, R3szchk, dboralib, ...),
    but when we start Import process we get an error in SAPSDIC package.
    The error looks like:
    *************************+
    (RTF) ########## WARNING ###########
    Without ORDER BY PRIMARY KEY the exported data may be unusable
    for some databases
    (EXP) INFO: table DDNTF will be exported with sorting
    (RSCP) ERROR: 'DDNTF' in UMGCCTL: Guess has illegal value.
    (RSCP) WARN: DDNTF: Missing in UMGSTAT,
    ERROR in nametab conversion
    conversion of nametab table 'DDNTF ':
    rscpMCStartTab: rc = 128
    (CNV) ERROR: data conversion failed. rc = 2
    (DB) INFO: disconnected from DB
    /usr/sap/CRT/SYS/exe/run/R3load: job finished with 1 error(s)
    /usr/sap/CRT/SYS/exe/run/R3load: END OF LOG: 20091229193651
    Any idea??
    Thanks in advance
    Regards

    Yes I executed this Report using Menu Tool Bar.
    The first execution gave me some errors regarding some tables existing at database level but  not in DDIC. We ignored this errrors because those tables are not important.
    But the last try, we decided solve this inconsistencies deleteting those tables at database level, repeting report RADUCNT (in this case no errors appeared) and repeting the export with the same error..
    Now, we have seen in Tabscript STATUS of SPUMG there are some errors regarding DDNTF inconsistency found at table control. Reason 4.  But this reason means that faillback codepage missing at table control for this table. But as far I know, I can't fill failbak field for this kind of table...
    So know I'm reseting SPUMG, and reestarting again..
    But I don't understand what is the problem
    In a few minutes I will you inform about results
    Meanwhile is anybody have any suggestion or idea... please let me know
    Thanks!

  • Issue with Data Load Table

    Hi All,
           i am facing issue with apex 4.2.4 ,using the  Data Load Table concept's and in this look up used the
          Where Clause option  ,it seems to be not working this where clause ,Please help me on this

    hi all,
        it looks this where clause not filter with 'N'  data ,Please help me ,how to solve this or help me on this

Maybe you are looking for