Plz help upgrade issue moving data from char type structure to non char typ

Hi Experts
plz help its very urgent
Data :workout(5000) .
FIELD-SYMBOLS : <FS_WORKOUT> TYPE ANY.  
workout = '         u' .
ASSIGN WORKOUT TO <FS_WORKOUT> CASTING TYPE C .
                  BAPISDITM = <FS_WORKOUT>.
i am getting dump after BAPISDITM = <FS_WORKOUT>.
i think i am getting the dump bcoz i am moving character type structure to non character type structure but i think with field symbols we can remove this issue thats y i used it but its not working plz help me
its very urgent
*dump is :*
Short text
    Data objects in Unicode programs cannot be converted.
What happened?
    Error in the ABAP Application Program
    The current ABAP program "ZSDR0009" had to be terminated because it has
    come across a statement that unfortunately cannot be executed.
How to correct the error
    Use only convertible operands "dst" and "src" for the statement
       "MOVE src TO dst"
    If the error occures in a non-modified SAP program, you may be able to
    find an interim solution in an SAP Note.
    If you have access to SAP Notes, carry out a search with the following
    keywords:
    "UC_OBJECTS_NOT_CONVERTIBLE" " "
    "ZSDR0009" or "ZSDR0009_I02"
    "USER_COMMAND"
thanx in advance

i got d solution in this thread
Hi all,
data: gv_line(6000) type c.
Bvbapkom = gv_line.
But i am getting the Error like : gv_line and Bvbapkom are not mutually convertable.
Note: Bvbapkom is a Structure
How do i solve this ?
Mahesh
KR  
Posts: 210
Registered: 11/24/06
Forum Points: 0 
  Re: gv_line and Bvbapkom are not mutually convertable.  
Posted: Nov 30, 2007 8:40 AM    in response to: KR         Reply 
Hi ,
i got the solution
ANSWER:
Field-symbols: <X_Bvbapkom> type x,
<X_gv_line> type x.
Assign: Bvbapkom to <X_Bvbapkom> casting,
gv_line to <X_gv_line> casting.
<X_Bvbapkom> = <X_gv_line>.
Nasaka Ramakris...  
Posts: 4
Registered: 1/19/08
Forum Points: 20 
  Re: gv_line and Bvbapkom are not mutually convertable.   
Posted: Jan 19, 2008 7:42 AM    in response to: KR         Reply 
Hi Check this answer.
ANSWER:
Field-symbols: <X_Bvbapkom> type x,
<X_gv_line> type x.
Assign: Bvbapkom to <X_Bvbapkom> casting,
gv_line to <X_gv_line> casting.
<X_Bvbapkom> = <X_gv_line>.

Similar Messages

  • Issue in Data from DSO to DSO Target with different Key

    Hello All,
    I am having Issue in Data from DSO to DSO Target with different Key
    Source DSO has Employee + Wage + Post numner as key and Target has Employee + Wage Type as key.
    DSO semantic grouping works like Group By clause in sql, is my understanding right ?
    Also if someone can explain this with a small example, it would be great.
    Many Thanks
    Krishna

    Dear, as explained earlier your issue has nothing to do with semantic grouping .
    Semantic grouping is only usefull when you have written a routine in the transformation for calculations and in error handling .
    Please go through this blog which explains very clearly the use of semantic grouping .
    http://scn.sap.com/community/data-warehousing/netweaver-bw/blog/2013/06/16/semantic-groups-in-dtp
    Now coming to your above question
    DSO 1
    Employee WageTyp Amount
    100          aa          200
    200          aa          200
    100          bb          400
    100          cc          300
    If we have semantic group as Employee .   If we have Employee as key of the target DSO and update type as summation .
    then target DSO will have
    Emp                Amount
    100                 700
    200                 200
    In this case Wage type will be the last record arriving from the data package . If the record 100  cc  300 is arrivng last then wage type will be cc .
    2) Case 2
    DSO 1
    Employee WageTyp Amount
    100          aa          200
    200          aa          200
    100          bb          400
    100          aa          300
    if we do Semantic grouping with Emp and Wage Type   If we have Employee and wage type as key of the target DSO and update type as summation .
    then target DSO will have
    Emp     Wage      Amount
    100          aa          500
    200          aa          200
    100          bb          400
    Hope this helps .

  • Moving Data from Normal table to History tables

    Hi All,
    I'm in the process of moving data from normal tables to
    History tables.
    It can be some sort of a procedure which should be a cron job running at night.
    My aim is to move data say 1.5 yrs or 2yrs old data to History tables.
    What aspects i need to check when moving data. And how can i write a procedure for this requirement.
    The schema is same in both the normal table and history table.
    It has to be a procedure based on particular field RCRE_DT.
    If the rcre_dt is above 2 yrs the data needs to be moved to HIS_<table>.
    I have to insert record in to HIS_table and simultaneously delete record from the normal table.
    This is in Production system and the tables are quite big.
    Pls do find enclosed the attached sample schema for Normal table and HIS_<table>.
    If i want to automate this script as a Cron job for similarly other History tables
    how am i to do it in a single procedure assuming the procedure for moving the data is the same procedure.
    Thanks for ur help in advance.
    SQL> DESC PXM_FLT;
    Name Null? Type
    RCRE_USER_ID NOT NULL VARCHAR2(15)
    RCRE_DT NOT NULL DATE
    LCHG_USER_ID VARCHAR2(15)
    LCHG_DT DATE
    AIRLINE_CD NOT NULL VARCHAR2(5)
    REF_ID NOT NULL VARCHAR2(12)
    BATCH_DT NOT NULL DATE
    CPY_NO NOT NULL NUMBER(2)
    ACCRUAL_STATUS NOT NULL VARCHAR2(1)
    FLT_DT NOT NULL DATE
    OPERATING_CARRIER_CD NOT NULL VARCHAR2(3)
    OPERATING_FLT_NO NOT NULL NUMBER(4)
    MKTING_CARRIER_CD VARCHAR2(3)
    MKTING_FLT_NO NUMBER(4)
    BOARD_PT NOT NULL VARCHAR2(5)
    OFF_PT NOT NULL VARCHAR2(5)
    AIR_CD_SHARE_IND VARCHAR2(1)
    UPLOAD_ERR_CD VARCHAR2(5)
    MID_PT1 VARCHAR2(5)
    MID_PT2 VARCHAR2(5)
    MID_PT3 VARCHAR2(5)
    MID_PT4 VARCHAR2(5)
    MID_PT5 VARCHAR2(5)
    PAX_TYPE VARCHAR2(3)
    PAY_PRINCIPLE VARCHAR2(1)
    SQL> DESC HIS_PXM_FLT;
    Name Null? Type
    RCRE_USER_ID NOT NULL VARCHAR2(15)
    RCRE_DT NOT NULL DATE
    LCHG_USER_ID VARCHAR2(15)
    LCHG_DT DATE
    AIRLINE_CD NOT NULL VARCHAR2(5)
    REF_ID NOT NULL VARCHAR2(12)
    BATCH_DT NOT NULL DATE
    CPY_NO NOT NULL NUMBER(2)
    ACCRUAL_STATUS NOT NULL VARCHAR2(1)
    FLT_DT NOT NULL DATE
    OPERATING_CARRIER_CD NOT NULL VARCHAR2(3)
    OPERATING_FLT_NO NOT NULL NUMBER(4)
    MKTING_CARRIER_CD VARCHAR2(3)
    MKTING_FLT_NO NUMBER(4)
    BOARD_PT NOT NULL VARCHAR2(5)
    OFF_PT NOT NULL VARCHAR2(5)
    AIR_CD_SHARE_IND VARCHAR2(1)
    UPLOAD_ERR_CD VARCHAR2(5)
    MID_PT1 VARCHAR2(5)
    MID_PT2 VARCHAR2(5)
    MID_PT3 VARCHAR2(5)
    MID_PT4 VARCHAR2(5)
    MID_PT5 VARCHAR2(5)
    PAX_TYPE VARCHAR2(3)
    PAY_PRINCIPLE VARCHAR2(1)

    Hi All,
    Thanks for ur valuable suggestion.But can u explain me bit more on this as i'm still confused about switching between partitoned tables and temporary table.Suppose if i have a table called PXM_FLT and an correspoding similar table named HIS_PXM_FLT.How can i do the partitioning shd i do the partitioning on the normal table or HIS_PXM_FLT.i do have a date field for me to partition based on range.Can u pls explain why shd i again create a temp.table.What's the purpose.Now the application is designed in such a way that old records have to be moved to HIS_PXM_FLT.can u pls highlight more on this.Your suggestions are greatly appreciated.As i'm relatively new to this partitioning technique i'm bit confused on how it works.But i came to understand Partitioning is a better operation than the normal insert or delte as it is more data intensive as millions of record need to be moved.Thanks for feedback and ur precious time.

  • Moving data from an array to a JTable

    Hi,
    I am facing some problems with handling array data to JTable.
    My objective is moving data from a 10X5 data array to a JTable.
    I am using the following code:
    String columnNames = { "Col 1", "Col 2", "Col 3", "Col 4", "Col 5" };
    for(int j=0; j<10; j++)
    data[ j ] = array1[ j ] + array2[ j ] + array3[ j ] + array4[ j ] + array5[ j ];
    JTable table = new JTable( data, coulmnNames);
    Can you please help me with the error in tha above code.My array1 is a String array and the other four arrays are of Integer data type.
    Thanks!!!

    hi,
    The actual error message is:
    [root@localhost java]# javac GUI.java
    GUI.java:144: incompatible types
    found : int
    required: java.lang.Object
    data [j] [1] = array1[j];
    ^
    GUI.java:145: incompatible types
    found : int
    required: java.lang.Object
    data [j] [2] = array2[j];
    ^
    GUI.java:146: incompatible types
    found : int
    required: java.lang.Object
    data [j] [3] = array3[j];
    ^
    GUI.java:147: incompatible types
    found : int
    required: java.lang.Object
    data [j] [4] = array4[j];
    ^
    4 errors
    My code segment with line no goes as follows:
    private static JTable createTable() {         
           String[] columnNames = { "Col 1", "Col 2", "Col 3", "Col 4"' "Col 5" };
             Object data [ ] [ ] = new Object [10] [5];
       for( int j=0; j<10; j++)
              data [j] [0] = array1[j]; ...................Line:144
              data [j] [1] = array2[j]; ..................Line 145
              data [j] [2] = array3[j]; ....................Line 146
              data [j] [3] = array4[j]; .................Line 147
              data [j] [4] = array5[j];
      final JTable table = new JTable(data, columnNames);
      return table;
    [code/]
    Array 1 is declared as String array and Array 2,3,4,5 are declared as Integer arrays.The error is coming from the Integer arrays only.
    Java version running on my machine is: 1.4.2
    Please help!!!
    Thanks!!!                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                               

  • Moving data from flat file

    Hi,
    I am moving data from flat file to oracle table. while populating the oracle table if I get any errors in flat file those errors should populate in ODI error table.
    Is this is possible? if yes. Could you please let me know the set up in ODI.

    CKM is the dedicated for checking the constraints while doing transformation. The constraints includes, PK,FK, conditions etc.,
    There are two types/ways of checking the constraints.
    Flow Control: Triggered CKM after data is loaded in to I$.
    Static Control : Triggered CKM after data is loaded in to target table.
    If you opt for any one the above ODI will create E$ and SNP_CHECK_TAB (summary table for logging the errors) and load the error records.
    ODI will also provide you an option of loading the corrected error records by RECYCLE ERROR feature. In this phase/run ODI will load ONLY the error records in to target table (assuming its been corrected/cleaned).
    **how to set up flow control could you please provide steps**
    **Appreciate your help**

  • Which CKM is used for moving data from Oracle to delimited file ?

    Hi All
    Please let me know Which CKM is used for moving data from Oracle to delimited file ?
    Also is there need of defining each columns before hand in target datastore. Cant ODI take it from the oracle table itself ?

    Addy,
    A CKM is a Check KM which is used to validate data and log errors. It is not going to assist you in data movement. You will need an LKM SQL to File append as answered in another thread.
    Assuming that you have a one to one mapping, to make things simpler you can duplicate the Oracle based model and create a file based model. This will take all the column definitions from the Oracle based model.
    Alternatively, you can also use an ODI tool odiSQLUnload to dump the data to a file
    HTH

  • Need help in fetching requested data from JSP

    Hello,
    I really need help in fecthing requested data from JSP to servlet. Can anyone assist me as soon
    as possible because I must finish my program by today.....( 20/02/2002).
    Thanks in advance.

    It is very likely that somebody can help you, if you say what your problem is. In fact somebody might already have helped you. What is your problem?

  • Help required to get data from Cash Mgt

    Hi
    I need help to get following data from cash managemment .
    Bank Name
    Branch
    Account#
    Date
    cheque#/Ref
    Amount Direct Debited by bank
    Amount Direct credeted by bank.
    Kindly help
    Regards

    Hello.
    See if this helps. When the Amount column shows a negative number it represents a Debit.
    select b.bank_account_name "Bank Account" ,
    bb.bank_branch_name "Branch",
    b.bank_account_num "Bank Acc Num",
    l.trx_date "Date",
    l.trx_type "Type",
    l.trx_text "Description",
    l.amount "Amount"
    from ce_statement_headers_all h,
    ce_statement_lines l,
    ap_bank_accounts_all b,
    ap_bank_branches bb
    where b.bank_account_id = h.bank_account_id
    and l.statement_header_id = h.statement_header_id
    and bb.bank_branch_id = b.bank_branch_id
    and b.bank_account_id = <bank_account_id>
    and h.statement_number = <'statement number'>
    Octavio

  • Error occurred moving data from the transformation RECON_SEQ_GEN: to the transformation FILTRANS.

    Hi,
    Today one my task in the DAC Etl failed and when I opened the DAC log file of it. I saw the error description as below
    Error while contacting Informatica server for getting workflow status for PLP_SRC_WH_AP_Invoice_Reconciliation
    Error Code = 36331:Unknown reason for error code 36331
    Since there is no much information in DAC log file about the failure, I opened the PLP_SRC_WH_AP_Invoice_Reconciliation session log file.
    The error description in the session log file says.
    There is an error in the port [NEXTVAL]: The default value for the port is set to: ERROR(<<Expression Error>> [ERROR]: transformation error
    ... nl:ERROR(u:'transformation error')).
    2014-01-14 08:00:45 : ERROR : (582188 | TRANSF_1_1_1) : (IS | Oracle_BI_DW_Server) : node01_ : TT_11023 : An error occurred converting data in the port [NEXTVAL]. The row was dropped.
    2014-01-14 08:00:45 : ERROR : (582188 | TRANSF_1_1_1) : (IS | Oracle_BI_DW_Server) : node01_ : TT_11021 : An error occurred moving data from the transformation RECON_SEQ_GEN: to the transformation FILTRANS.
    2014-01-14 08:00:45 : ERROR : (582188 | TRANSF_1_1_1) : (IS | Oracle_BI_DW_Server) : node01 : CMN_1086 : RECON_SEQ_GEN: Number of errors exceeded threshold [1].
    2014-01-14 08:00:45 : INFO : (582188 | TRANSF_1_1_1) : (IS | Oracle_BI_DW_Server) : node01_ : CMN_1053 : : Rowdata: ( RowType=-4(error but dropped due to abort) Src Rowid=1186281 Targ Rowid=1186281
      NEXTVAL (ROW_WID:Int:): "(NULL)"
    I am not sure of this error and I have never seen before. Can any one share some ideas on how to tackle this error and make the task execute successfully.
    Many Thanks.

    Hi,
    fist check weather any special characteristics if not
    check in data source under this we have fields tab check the format of a particular field format internal/external/check u choose internal format, if any check routine once
    use Semantic Groups in the DTP.
    Try  it
    Thanku
    lokeeshM
    Edited by: lmedaSAP_BI on Oct 20, 2010 6:44 AM

  • How to read data from a CLUSTER STRUCTURE not cluster table.

    Hi,
    how to read data from a CLUSTER STRUCTURE not cluster table.
    regards,
    Usha.

    Hello,
    A structre doesnt contain data.. so u cannot read from it. U need to find out table of that structure and read data from it.
    Regards,
    Mansi.

  • How to read data from the Deep structure.

    Hi
    How to get from the deep structure. Means i have one table in that having onother structure. i need to read data from that inner structure.
    Regards
    Reddy

    Hi,
    you can access in the way u use for the normal structure, u should only consider a deep structure is a table without headerline.
    TABLES: BKPF, BSEG.
    TYPES: TY_ITEM TYPE TABLE OF BSEG.
    DATA:   BEGIN OF W_DOCUMENT,
                    HEADER TYPE BKPF,
                    ITEM       TYPE BSEG,
                 END    OF W_DOCUMENT.
    DATA: T_DOCUMENTS LIKE STANDARD TABLE OF W_DOCUMENT.
    Insert the data:
    SELECT * FROM BKPF WHERE ....
       W_DOCUMENT-HEADER = BKPF.
       SELECT * FROM BSEG INTO TABLE W_DOCUMENT-ITEM
                                       WHERE BUKRS = BKPF-BUKRS
                                            AND BELNR  = BKPF-BELNR
                                            AND GJAHR  = BKPF-GJAHR.
       APPEND W_DOCUMENT TO T_DOCUMENTS.
    ENDSELECT.
    Read the data:
    LOOP AT T_DOCUMENTS INTO W_DOCMENT.
    Header data
        WRITE: / W_DOCUMENT-HEADER-BUKRS,
                      W_DOCUMENT-HEADER-BELNR,
                      W_DOCUMENT-HEADER-GJAHR.
    Item data
        LOOP AT W_DOCUMENT-ITEM INTO BSEG.
             WRITE: / BSEG-BUZEI,
                           BSEG-WRBTR CURRENCY W_DOCUMENT-HEADER-WAERS.
        ENDLOOP.
    ENDLOOP.
    Regards,
    Padmam.

  • Plz help production issue(smartform)

    hi experts plz help its a production issue
    i have a box in my smartform in which in one line i can only show 50 characters .and i have a text of 70 characters but i only have to show 50 characters from 70 characters .
    now wat i did i stored 70 characters variable into 50 char variable abd displaying it .
    but the issue is
    if i want to print .
    hi this is sdn and here u can get all the information about sap .
    now lets say
    i of information is the 46th character of my line
    now what is happening it is coming like
    hi this is sdn and here u can get all the
    infor
    its showing 50 character only but if word is not completed it is printing it in next line.
    it should print like
    hi this is sdn and here u can get all the infor
    plz help how to do it
    thanx in advance .

    code i am using is ::::::::::::::::::::::::::::::::::::::::::::::::
    data : TEXT_3I56(50) type c .
    move is_bil_invoice-hd_gen-bil_number TO textname.
    CALL FUNCTION 'READ_TEXT'
    EXPORTING
    CLIENT                         = SY-MANDT
    id                            = '0002'
    language                      = is_nast-spras
    name                          = textname
    object                        = 'VBBK'
      ARCHIVE_HANDLE                = 0
      LOCAL_CAT                     = ' '
    IMPORTING
      HEADER                        =
    tables
    lines                         = t_lines_3156
    EXCEPTIONS
    ID                            = 1
    LANGUAGE                      = 2
    NAME                          = 3
    NOT_FOUND                     = 4
    OBJECT                        = 5
    REFERENCE_CHECK               = 6
    WRONG_ACCESS_TO_ARCHIVE       = 7
    OTHERS                        = 8.
    IF sy-subrc <> 0.
    MESSAGE ID SY-MSGID TYPE SY-MSGTY NUMBER SY-MSGNO
            WITH SY-MSGV1 SY-MSGV2 SY-MSGV3 SY-MSGV4.
    ENDIF.
    *describe table t_lines lines n.
    *if not n is initial.
    *bank_note = 'X'.
    *endif.
    READ TABLE T_LINES_3156 INDEX 1 .
    IF SY-SUBRC = 0 .
    TEXT_3I56 = T_LINES_3156-TDLINE .
    ENDIF.

  • What are required Oracle products for moving data from IBM IMS/DB(mainframe) to Oracle environment?

    I am z/OS system programmer, our company is using IMS as its main OLTP database. We are investigating moving data off the mainframe for data warehousing and online fraud detection. One option is using IBM InfoSphere CDC and DB2, another option is using IMS connect and writing our own program, I am wondering what is the oracle solution for this kind of issue?
    I am not oracle technician but I googled and find out Oracle has some product like Oracle Legacy Adapters, OracleAS CDC Adapter and Oracle Connect on z/OS, however I didn't find them in Oracle site(https://edelivery.oracle.com/), I don't know whether these products are deprecated or not?!
    I would very much appreciate any help or guidance you are able to give me

    Thank you for responding.
    I've considered dumping the data into a flat file and using SQL*Loader to import as you suggest but this would require some scripting on a per-table basis. Again: all I want to do is to copy the contents of a table from one database to another. I do not think I should have to resort to creating my own dump and load scripts in order to do that. However I agree with you that this type of solution may be my final solution.
    I've tried the db link solution. It was just a slow as the 'imp' solution for some reason. Don't know why. The tables are rather large (3 tables of a few Gb each) and therefore require intermediate commits when loaded. Otherwise the rollback segment will run out of space. So the 'db link solution' is really a PL/SQL script with a commit for each x records.
    I think Oracle is making it a bit difficult for me to copy the contents of a table from one database to another and to do it efficiently. Perhaps I'm missing something here?

  • Error while moving Data from fs to its corresponding Structure

    Hi Gurus,
    I am getting a run time error  (Operands are not Convertible) while executing the below statement:
    MOVE <fs> TO  ls_struc.
    [ Over here: ls_struc is TYPE XYZ (a custom structure containg 7 fields, out of which one field is type Packed decimal, which is why I get a run time error)  AND
                         <fs> is a field symbol which is containing the data for all fields in the structure XYZ ]
    If I change the data type of packed decimal field in the structure: 'LS_STRUC' to 'char ' type, then the above query works.
    Could anyone help in resolving this issue.
    Thanks
    Gaurav Verma

    Hi Gverma09,
    Good!
    you can do a bit simpler without structdescr and without handle:
    CREATE DATA e_dref TYPE (im_stru_name).
    or even
    CREATE DATA e_dref like <fs>.
    Access  as you do now, eventually need second field-symbol.
    ASSIGN lv_dataref->* TO <fs2>
    The ABAP RTTS Run Time Type Services CL_ABAB_xyzDESCR are especially useful if you need to know all field properties of a structure or if you want to create structures and tables dynamically. In this case, CREATE DATA has the required features.
    Regards,
    Clemens

  • Help needed in  extracting data from PCD tables

    Hi Friends
    I Have a requiremnt for creating custom portal activity report ,even though
    we have  standard report, the extraced data will be used to create bw reports later.
    my part is to find a way to extract the data from PCD tables for creating
    custom portal activity reports
    i have selected the following  tables for the data extraction
    WCR_USERSTAT,WCR_WEBCONTENTSTAT,WCR_USERFIRSTLOGON,
    WCR_USERPAGEUSAGE.
    My questions are
    1.Did i select the Exact PCD tables?
    2.Can i use UME api  for  accessing the data from those tables?
    3.can i use  the data extracted  from PCD tables in JSPdynpage  or
    webdynpro apps?
    4.can i Querry  the  PCD tables from  JSPDynpage or Webdynpro
    Please help me in finding a solution for this
    Thanks
    Ashok Battula

    Hi daniel
    Can u tell  me weather i can develop the following  custom reports from those WCR tables
         Report Type
    1     Logins
          - Unique Count
          - Total Count
          - Most Active Users (by Partner Name)
          - Most Active Users (by Contact Name)
          - Entry Point (by page name)
          - Session Time
          - Hourly Traffic Analysis
    2     Login Failures
          - Total Count
          - Count by error message
          - Credentials Entered (by user name and password)
    3     Content Views (by File Name)
          - Unique Count
          - Total Count
          - Most requested Files
          - Most requested Pages
          - File Not Found
    4     Downloads (by File Name)
          - Unique Count
          - Total Count
          - Most requested Files
          - File Not Found
    5     Portal Administration
          - Site Content (by file name)
          - Site Content (by page name)
          - Latest Content (by file name)
          - Expired Content (by file name)
          - Subscriptions Count (by file name)
    6     Login History (by Partner, Contact Name)
          - No Login
          - First Login
          - Duration between registration and first login
          - Most Recent Login
          - Average Number of Logins
    plz  help me in find ing a way
    thanks
    ashok

Maybe you are looking for