Apps team uploading 114505 kbof text file,too many archives getting

the upload is going on ,too many are getting generated in archivelog location
i have increased 200gb to 900gb
please let me know how to know for 14505 kb of dataupload how many archives gettting generated
can we estimate before dataupload how many archives can generate
can you please explain clearly
thanks

Hi,
There is no specific method to calculate the number of archived log files to be generated for the particular load.
But, we may do some rough calculation by looking at the rate of archived log generation.
Check the following things,
1) How frequently the logs are getting generated and the size of the files.
2) How many data has been loaded so far.
Based on the above input, you may get some idea on number of ARCHIVED LOG files to be generated during the rest of the data load.
Also, you may do the following to avoid keep adding space to archive destination.
1) Wite a UNIX shell script to BACKUP the ARCHIVED LOG FILES and DELETE the files once backuped up; which should run for every 10 minutes.
Thanks.

Similar Messages

  • Uploading comma delimited text file

    Dear Experts,
    I want to upload comma delimited text file in sap and get that values into internal table
    Can any 1 provide a sample code ??
    thanx in advance.

    Hi Suchitra ,
    There is FM GUI_UPLOAD which will help you in this.
    Its has got one parameter calles has_field_separator . Here you can pass comma. So that this will take care while uploading the data in Internal table.
    CALL FUNCTION 'GUI_UPLOAD'
        EXPORTING
          filename                      = wf_file
          filetype                      = 'ASC'
          has_field_separator           = ','
    *   HEADER_LENGTH                 = 0
    *   READ_BY_LINE                  = 'X'
    *   DAT_MODE                      = ' '
    *   CODEPAGE                      = ' '
    *   IGNORE_CERR                   = ABAP_TRUE
    *   REPLACEMENT                   = '#'
    *   CHECK_BOM                     = ' '
    * IMPORTING
    *   FILELENGTH                    =
    *   HEADER                        =
        TABLES
          data_tab                      = it_data
       EXCEPTIONS
         file_open_error               = 1
         file_read_error               = 2
         no_batch                      = 3
         gui_refuse_filetransfer       = 4
         invalid_type                  = 5
         no_authority                  = 6
         unknown_error                 = 7
         bad_data_format               = 8
         header_not_allowed            = 9
         separator_not_allowed         = 10
         header_too_long               = 11
         unknown_dp_error              = 12
         access_denied                 = 13
         dp_out_of_memory              = 14
         disk_full                     = 15
         dp_timeout                    = 16
         OTHERS                        = 17
    Hope this will help you.
    Regards,
    Nikhil

  • Air App will not read local text file using openAsync/readUTFBytes on user (non-admin) mode

    I am running an Air App I did for the desktop, from the actual installed executable already deployed in the machine (Not from Flash Pro / Flex dev. environment). For some reason the app will not read a text file stored in the same application folder unless I run my app as administrator from the OS.
    When I run the app as admin, or within the development environment it works fine. Maybe this is related to some security issue? I read the adobe air documentation, and this should work...
    I am using openAsync/readUTFBytes on user as shown here:
    var continueGamesConnection:FileStream();
    var continueFile:File = new File(File.applicationDirectory.resolvePath("continueGames.txt").nativePath.toString()); 
    continueGamesConnection.addEventListener(Event.COMPLETE, openSavedGames);
    continueGamesConnection.openAsync(continueFile, FileMode.UPDATE); 
    function openSavedGames(event:Event):void
         continueGamesConnection.removeEventListener(Event.COMPLETE, openSavedGames);
         var content:URLVariables = new URLVariables();
         var loadedContent:String = new String();
         loadedContent = continueGamesConnection.readUTFBytes(continueGamesConnection.bytesAvailable);
         content.decode(loadedContent); 
         variableX = content. variableX
         //etc, etc. 
         continueGamesConnection.close();
    By the way, I have also, tried using FileMode.READ, and others, and it still gives me the same problem. Only works if ran on admin mode or from the dev. environment.
    It's very frustrating, I tried reading other posts without any luck... What solutions do people use for this kind of problem?
    I have seen that you can set the app to run as admin somehow, and I guess that could work. However, this should work just fine, since it doesn't seem to violate any of the security APIs of Air. Seems like an overkill. But even so, how do I do that?
    You help is greatly appreciated!

    Thanks kglad.com. I will try this and see if it works. Can you check my code a bit to see if it's right?
    var continueFile:File = new File(File.applicationStorageDirectory.resolvePath("savedgames/continueGames.txt").nativePath.toString());
    Does this look right to work across all desktop OS?

  • I have a problem with my mac  i have too many  archives and programs and i want to delete all files and start at the begin my mac with out do format

    i have a problem with my mac  i have too many  archives and programs and i want to delete all files and start at the begin my mac with out do format

    I do not recommend reformatting your harddrive and reloading your software.  This is a Windows thing. On an older mac it may be difficult to find all your software again.
    Best to have greater than 2gig of free space.  Many posters to these forums state that you need much more free space: 5 gig to 10 gig or 10 percent of you hd size.
    (0)
    Be careful when deleting files. A lot of people have trashed their system when deleting things. Place things in trash. Reboot & run your applications. Empty trash.
    Go after large files that you have created & know what they are.  Do not delete small files that are in a folder you do not know what the folder is for. Anything that is less than a megabyte is a small file these days.
    (1)
    Run
    OmniDiskSweeper
    "The simple, fast way to save disk space"
    OmniDiskSweeper is now free!
    http://www.omnigroup.com/applications/omnidisksweeper/download/
    This will give you a list of files and folders sorted by size. Go after things you know that are big.
    (2)
    These pages have some hints on freeing up space:
    http://thexlab.com/faqs/freeingspace.html
    http://www.macmaps.com/diskfull.html
    (3)
    Buy an external firewire harddrive.
    For a PPC computer, I recommend a firewire drive.
    Has everything interface:
    FireWire 800/400 + USB2, + eSATA 'Quad Interface'
    save a little money interface:
    FireWire 400 + USB 2.0
    This web page lists both external harddrive types. You may need to scroll to the right to see both.
    http://eshop.macsales.com/shop/firewire/1394/USB/EliteAL/eSATA_FW800_FW400_USB
    (4)
    Buy a flash card.

  • Create procedure is generating too many archive logs

    Hi
    The following procedure was run on one of our databases and it hung since there were too many archive logs being generated.
    What would be the answer? The db must remain in archivelog mode.
    I understand the nologging concept, but as I know this applies to creating tables, views, indexes and tablespaces. This script is creating procedure.
    CREATE OR REPLACE PROCEDURE APPS.Dfc_Payroll_Dw_Prc(Errbuf OUT VARCHAR2, Retcode OUT NUMBER
    ,P_GRE NUMBER
    ,P_SDATE VARCHAR2
    ,P_EDATE VARCHAR2
    ,P_ssn VARCHAR2
    ) IS
    CURSOR MainCsr IS
    SELECT DISTINCT
    PPF.NATIONAL_IDENTIFIER SSN
    ,ppf.full_name FULL_NAME
    ,ppa.effective_date Pay_date
    ,ppa.DATE_EARNED period_end
    ,pet.ELEMENT_NAME
    ,SUM(TO_NUMBER(prv.result_value)) VALOR
    ,PET.ELEMENT_INFORMATION_CATEGORY
    ,PET.CLASSIFICATION_ID
    ,PET.ELEMENT_INFORMATION1
    ,pet.ELEMENT_TYPE_ID
    ,paa.tax_unit_id
    ,PAf.ASSIGNMENT_ID ASSG_ID
    ,paf.ORGANIZATION_ID
    FROM
    pay_element_classifications pec
    , pay_element_types_f pet
    , pay_input_values_f piv
    , pay_run_result_values prv
    , pay_run_results prr
    , pay_assignment_actions paa
    , pay_payroll_actions ppa
    , APPS.pay_all_payrolls_f pap
    ,Per_Assignments_f paf
    ,per_people_f ppf
    WHERE
    ppa.effective_date BETWEEN TO_DATE(p_sdate) AND TO_DATE(p_edate)
    AND ppa.payroll_id = pap.payroll_id
    AND paa.tax_unit_id = NVL(p_GRE, paa.tax_unit_id)
    AND ppa.payroll_action_id = paa.payroll_action_id
    AND paa.action_status = 'C'
    AND ppa.action_type IN ('Q', 'R', 'V', 'B', 'I')
    AND ppa.action_status = 'C'
    --AND PEC.CLASSIFICATION_NAME IN ('Earnings','Alien/Expat Earnings','Supplemental Earnings','Imputed Earnings','Non-payroll Payments')
    AND paa.assignment_action_id = prr.assignment_action_id
    AND prr.run_result_id = prv.run_result_id
    AND prv.input_value_id = piv.input_value_id
    AND piv.name = 'Pay Value'
    AND piv.element_type_id = pet.element_type_id
    AND pet.element_type_id = prr.element_type_id
    AND pet.classification_id = pec.classification_id
    AND pec.non_payments_flag = 'N'
    AND prv.result_value <> '0'
    --AND( PET.ELEMENT_INFORMATION_CATEGORY LIKE '%EARNINGS'
    -- OR PET.element_type_id IN (1425, 1428, 1438, 1441, 1444, 1443) )
    AND NVL(PPA.DATE_EARNED, PPA.EFFECTIVE_DATE) BETWEEN PET.EFFECTIVE_START_DATE AND PET.EFFECTIVE_END_DATE
    AND NVL(PPA.DATE_EARNED, PPA.EFFECTIVE_DATE) BETWEEN PIV.EFFECTIVE_START_DATE AND PIV.EFFECTIVE_END_DATE --dcc
    AND NVL(PPA.DATE_EARNED, PPA.EFFECTIVE_DATE) BETWEEN Pap.EFFECTIVE_START_DATE AND Pap.EFFECTIVE_END_DATE --dcc
    AND paf.ASSIGNMENT_ID = paa.ASSIGNMENT_ID
    AND ppf.NATIONAL_IDENTIFIER = NVL(p_ssn, ppf.NATIONAL_IDENTIFIER)
    ------------------------------------------------------------------TO get emp.
    AND ppf.person_id = paf.person_id
    AND NVL(PPA.DATE_EARNED, PPA.EFFECTIVE_DATE) BETWEEN ppf.EFFECTIVE_START_DATE AND ppf.EFFECTIVE_END_DATE
    ------------------------------------------------------------------TO get emp. ASSIGNMENT
    --AND paf.assignment_status_type_id NOT IN (7,3)
    AND NVL(PPA.DATE_EARNED, PPA.EFFECTIVE_DATE) BETWEEN paf.effective_start_date AND paf.effective_end_date
    GROUP BY PPF.NATIONAL_IDENTIFIER
    ,ppf.full_name
    ,ppa.effective_date
    ,ppa.DATE_EARNED
    ,pet.ELEMENT_NAME
    ,PET.ELEMENT_INFORMATION_CATEGORY
    ,PET.CLASSIFICATION_ID
    ,PET.ELEMENT_INFORMATION1
    ,pet.ELEMENT_TYPE_ID
    ,paa.tax_unit_id
    ,PAF.ASSIGNMENT_ID
    ,paf.ORGANIZATION_ID
    BEGIN
    DELETE cust.DFC_PAYROLL_DW
    WHERE PAY_DATE BETWEEN TO_DATE(p_sdate) AND TO_DATE(p_edate)
    AND tax_unit_id = NVL(p_GRE, tax_unit_id)
    AND ssn = NVL(p_ssn, ssn)
    COMMIT;
    FOR V_REC IN MainCsr LOOP
    INSERT INTO cust.DFC_PAYROLL_DW(SSN, FULL_NAME, PAY_DATE, PERIOD_END, ELEMENT_NAME, ELEMENT_INFORMATION_CATEGORY, CLASSIFICATION_ID, ELEMENT_INFORMATION1, VALOR, TAX_UNIT_ID, ASSG_ID,ELEMENT_TYPE_ID,ORGANIZATION_ID)
    VALUES(V_REC.SSN,V_REC.FULL_NAME,v_rec.PAY_DATE,V_REC.PERIOD_END,V_REC.ELEMENT_NAME,V_REC.ELEMENT_INFORMATION_CATEGORY, V_REC.CLASSIFICATION_ID, V_REC.ELEMENT_INFORMATION1, V_REC.VALOR,V_REC.TAX_UNIT_ID,V_REC.ASSG_ID, v_rec.ELEMENT_TYPE_ID, v_rec.ORGANIZATION_ID);
    COMMIT;
    END LOOP;
    END ;
    So, how could I assist our developer with this, so that she can run it again without it generating a ton of logs ? ?
    Thanks
    Oracle 9.2.0.5
    AIX 5.2

    The amount of redo generated is a direct function of how much data is changing. If you insert 'x' number of rows, you are going to generate 'y' mbytes of redo. If your procedure is destined to insert 1000 rows, then it is destined to create a certain amount of redo. Period.
    I would question the <i>performance</i> of the procedure shown ... using a cursor loop with a commit after every row is going to be a slug on performance but that doesn't change the fact 'x' inserts will always generate 'y' redo.

  • Too many Archive Log

    Hi,
    I have a problem on a production server: our support partner told me that we are creating too many Archive Log files. They are not sure at all of what is happening.
    Please, do you know if there is any System View which we could obtain more information about Archive Logs usage?
    Thanks in advance.
    Best Regards,
    Joan Padilla

    Hi Joan,
    The sensible number of archive logs is determined by your backup and recovery strategy. You can delete them, once you have back-upped the database and the archivelogs. Further info in the Backup and Recovery Concepts manual.
    You can find information on them in v$archived_log.
    If you decide to delete them without having read the aforementioned manual, I will pray for you you don't suffer a hard disk crash.
    You may want to review the relationship with your support partner too.
    Their advice is not exactly professional.
    Sybrand Bakker
    Senior Oracle DBA

  • System I/O and Too Many Archive Logs

    Hi all,
    This is frustrating me. Our production database began to produce too many archived redo logs instantly --again. This happened before; two months ago our database was producing too many archive logs; just then we began get async i/o errors, we consulted a DBA and he restarted the database server telling us that it was caused by the system(???).
    But after this restart the amount of archive logs decreased drastically. I was deleting the logs by hand(350 gb DB 300 gb arch area) and after this the archive logs never exceeded 10% of the 300gb archive area. Right now the logs are increasing 1%(3 GB) per 7-8 mins which is too many.
    I checked from Enterprise Manager, System I/O graph is continous and the details show processes like ARC0, ARC1, LGWR(log file sequential read, db file parallel write are the most active ones) . Also Phsycal Reads are very inconsistent and can exceed 30000 KB at times. Undo tablespace is full nearly all of the time causing ORA-01555.
    The above symptoms have all began today. The database is closed at 3:00 am to take offline backup and opened at 6:00 am everyday.
    Nothing has changed on the database(9.2.0.8), applications(11.5.10.2) or OS(AIX 5.3).
    What is the reason of this most senseless behaviour? Please help me.
    Thanks in advance.
    Regards.
    Burak

    Selam Burak,
    High number of archive logs are being created because you may have massive redo creation on your database. Do you have an application that updates, deletes or inserts into any kind of table?
    What is written in the alert.log file?
    Do you have the undo tablespace with the guarentee retention option btw?
    Have you ever checked the log file switch sequency map?
    Please use below SQL to detirme the switch frequency;
    SELECT * FROM (
    SELECT * FROM (
    SELECT   TO_CHAR(FIRST_TIME, 'DD/MM') AS "DAY"
    +, TO_NUMBER(SUM(DECODE(TO_CHAR(FIRST_TIME, 'HH24'), '00', 1, 0)), '999') "00:00"+
    +, TO_NUMBER(SUM(DECODE(TO_CHAR(FIRST_TIME, 'HH24'), '01', 1, 0)), '999') "01:00"+
    +, TO_NUMBER(SUM(DECODE(TO_CHAR(FIRST_TIME, 'HH24'), '02', 1, 0)), '999') "02:00"+
    +, TO_NUMBER(SUM(DECODE(TO_CHAR(FIRST_TIME, 'HH24'), '03', 1, 0)), '999') "03:00"+
    +, TO_NUMBER(SUM(DECODE(TO_CHAR(FIRST_TIME, 'HH24'), '04', 1, 0)), '999') "04:00"+
    +, TO_NUMBER(SUM(DECODE(TO_CHAR(FIRST_TIME, 'HH24'), '05', 1, 0)), '999') "05:00"+
    +, TO_NUMBER(SUM(DECODE(TO_CHAR(FIRST_TIME, 'HH24'), '06', 1, 0)), '999') "06:00"+
    +, TO_NUMBER(SUM(DECODE(TO_CHAR(FIRST_TIME, 'HH24'), '07', 1, 0)), '999') "07:00"+
    +, TO_NUMBER(SUM(DECODE(TO_CHAR(FIRST_TIME, 'HH24'), '08', 1, 0)), '999') "08:00"+
    +, TO_NUMBER(SUM(DECODE(TO_CHAR(FIRST_TIME, 'HH24'), '09', 1, 0)), '999') "09:00"+
    +, TO_NUMBER(SUM(DECODE(TO_CHAR(FIRST_TIME, 'HH24'), '10', 1, 0)), '999') "10:00"+
    +, TO_NUMBER(SUM(DECODE(TO_CHAR(FIRST_TIME, 'HH24'), '11', 1, 0)), '999') "11:00"+
    +, TO_NUMBER(SUM(DECODE(TO_CHAR(FIRST_TIME, 'HH24'), '12', 1, 0)), '999') "12:00"+
    +, TO_NUMBER(SUM(DECODE(TO_CHAR(FIRST_TIME, 'HH24'), '13', 1, 0)), '999') "13:00"+
    +, TO_NUMBER(SUM(DECODE(TO_CHAR(FIRST_TIME, 'HH24'), '14', 1, 0)), '999') "14:00"+
    +, TO_NUMBER(SUM(DECODE(TO_CHAR(FIRST_TIME, 'HH24'), '15', 1, 0)), '999') "15:00"+
    +, TO_NUMBER(SUM(DECODE(TO_CHAR(FIRST_TIME, 'HH24'), '16', 1, 0)), '999') "16:00"+
    +, TO_NUMBER(SUM(DECODE(TO_CHAR(FIRST_TIME, 'HH24'), '17', 1, 0)), '999') "17:00"+
    +, TO_NUMBER(SUM(DECODE(TO_CHAR(FIRST_TIME, 'HH24'), '18', 1, 0)), '999') "18:00"+
    +, TO_NUMBER(SUM(DECODE(TO_CHAR(FIRST_TIME, 'HH24'), '19', 1, 0)), '999') "19:00"+
    +, TO_NUMBER(SUM(DECODE(TO_CHAR(FIRST_TIME, 'HH24'), '20', 1, 0)), '999') "20:00"+
    +, TO_NUMBER(SUM(DECODE(TO_CHAR(FIRST_TIME, 'HH24'), '21', 1, 0)), '999') "21:00"+
    +, TO_NUMBER(SUM(DECODE(TO_CHAR(FIRST_TIME, 'HH24'), '22', 1, 0)), '999') "22:00"+
    +, TO_NUMBER(SUM(DECODE(TO_CHAR(FIRST_TIME, 'HH24'), '23', 1, 0)), '999') "23:00"+
    FROM V$LOG_HISTORY
    WHERE extract(year FROM FIRST_TIME) = extract(year FROM sysdate)
    GROUP BY TO_CHAR(FIRST_TIME, 'DD/MM')
    +) ORDER BY TO_DATE(extract(year FROM sysdate) || DAY, 'YYYY DD/MM') DESC+
    +) WHERE ROWNUM < 8+
    Ogan

  • Unable to Upload data from text file into BEx Analyzer selection screen

    Hi,
    No response from BEx Analyzer when I am trying to upload around 40,000 material from text file into BEx Analyzer selection screen using "Upload selections" options. But I am able to upload only 10,000 material from text file.  I never faced same kind of issue when I am using BEx Analyzer 3.x.  Please let me know I have to change any settings related to BEx or any other.
    Thanks
    Sri Krishna Ponnada.

    Hello
    It seems you are reaching the .NET memory limitation informed in note 1040454.
    Because 3.5 does not use .NET it can work that.
    Regards,
    Ricardo

  • Error while uploading script form text file to SAP system

    hi
      i down loaded a SAP Script from production server to local test file using RSTXSCRP
    it get downloaded successfully but when i am uploading it to development server it is giving ERROR as :
    Transport object FORM ZIV_MEDRUCK_PO is being processed        
    Language vector used: DE                                       
    Original language was set to E                                 
    ERROR: Save with SAVE_TEXT(FORM ZIV_MEDRUCK_PO SAP DEF ) failed
    Text ID DEF 1 for text object FORM 2 does not exist            
    001 fatal error(s) occurred during import                      
    ERROR: Invalid start marker: instead of H                      
    ERROR: Invalid start marker: instead of H                     
    please help .

    Hi ,
    The text used in the previosu script is not in language en or proabably the standard text is not existent at the application server .
    Create the same text first then try uploading it it shold work.
    Regards,
    Nisrin.

  • Uploading many text file into many table

    hi there,
    my problem is that i have lots of text files (about 360 files), what i need is to upload them to many tables already created in oracle, as far as i know, i can upload only one file per command using the SQL loader, any help would be appreciated.
    thanks in advance.
    i am using oracle 10G and Windows XP SP3
    Edited by: user5724824 on Nov 20, 2009 4:36 AM

    Hello user5724824.
    Besides what drop.any and Toni Lazarin have provided, the only thing I could add would be the option to load one file to multiple tables, reading the file only once. Using Toni's example from another thread and a slight revision
    LOAD DATA
      INFILE mon200601.dat
      INFILE mon200602.dat
      INFILE mon200603.dat
      APPEND
      INTO TABLE emp1
      ( empno  POSITION(1:4)   INTEGER EXTERNAL,
        ename  POSITION(6:15)  CHAR,
        deptno POSITION(17:18) CHAR,
        mgr    POSITION(20:23) INTEGER EXTERNAL
      INTO TABLE emp2
      ( empno  POSITION(1:4)   INTEGER EXTERNAL,
        ename  POSITION(6:15)  CHAR,
        deptno POSITION(17:18) CHAR,
        mgr    POSITION(20:23) INTEGER EXTERNAL
      )If you want specific code to dynamically create all of your control files and execution of sqlldr for each created control file, a small example of table names and file names and their relationships that you'd like to be used would get me started. Include a three line sample example control file, that way we know whether you are using fixed or delimited data (this is significant).
    Hope this helps,
    Luke

  • FM to upload TAB DELIMITED TEXT file into Internal table.

    Hello Friends,
    The FM 'ALSM_EXCEL_TO_INTERNAL_TABLE' is used to upload EXCEL file into a Internal table.
    Is there any FM which performs the simillar operation on TAB DELIMITED TEXT FILE.
    Thanks in advance!
    Ashish

    Hi,
    To upload text file with tab delimated you can use FM
    GUI_OPLOAD.
    In this function you have put X in the field HAS_FIELD_SEPARATOR.
    Regards,
    Sujit

  • Upload and plot text files

    Hi
    I would like to upload and plot as overlays several text files in labview.
    Uploading and ploting one file is not a problem but I want to load several and see them overlayed in a single plot.  Presumably I need a looping routine that uploads each file independently putting the data in an array, then plots each of the arrays to the x-y graph.  I just can't figure a way of doing it.  Can anybody help?

    Hi.
    Here is a small example (LV7.1)
    Message Edited by EVS on 09-05-2005 05:01 PM
    Jack
    Win XP
    LabVIEW 6.1, 7.0, 7.1, LabWindows/ CVI 7.1
    Let us speek Russian
    Attachments:
    Add PlotXY.vi ‏56 KB
    Clip_7.jpg ‏64 KB

  • Upload time data text file in ztable for HR module

    dear all SAP guru's
    text file is like that
    260807300001
    270815480001
    270812210002.
    1st  4 char is date&month .
    2nd 4 char is time.
    3rd 4 char is employee no.
    i want to upload this file in z table.
    when i m uploading the time is not uploading correctly.my condition is like that
    already 1 condition is in Z table after that when i upload 2nd entry that time what 1st time is  time in and 2nd is  time out.
    coding as fallows . give any solution for that.
    LOOP AT it_out.
       CONCATENATE sy-datum0(4) it_out-text2(2) it_out-text+0(2) INTO itab1-val1.
       CONCATENATE it_out-text+4(4) '00' INTO itab1-val2.
       MOVE it_out-text+8(4) TO itab1-val3.
       APPEND itab1.
       CLEAR itab1.
    ENDLOOP.
    SELECT *
             FROM ztimedata
             INTO TABLE it_ztimedata.
    DESCRIBE TABLE it_ztimedata LINES cnt.
    LOOP AT it_out.
      CLEAR cnt.
        LOOP AT itab1.
          itab1-date1 = itab1-val1.
          itab1-timein = itab1-val2.
          itab1-pernr = itab1-val3.
            cnt = cnt + 1.
          itab1-counter = cnt.
          MODIFY itab1.
        ENDLOOP.
    ENDLOOP.
    SORT itab1 BY pernr date1 timein.
    DATA cnt1 TYPE i.
    DATA p_date LIKE sy-datum.
      SELECT SINGLE  * FROM ztimedata
                     INTO CORRESPONDING FIELDS OF  it_ztimedata.
    LOOP AT itab1.
              ztimedata-counter    = itab1-counter.
              ztimedata-date1      = itab1-val1.
              ztimedata-pernr      = itab1-val3.
              ztimedata-shift      = itab1-shift.
              ztimedata-attendance = itab1-attendance.
              CLEAR IT_ZTIMEDATA-TIMEIN.
           IF ztimedata-timein IS INITIAL.
              it_ztimedata-timein     = itab1-val2.
              ztimedata-timein     = itab1-val2.
              CLEAR it_ztimedata-timeout .
           ELSEIF ztimedata-timein IS NOT INITIAL.
              it_ztimedata-timeout     = itab1-val2.
              ztimedata-timeout    = itab1-val2.
           ENDIF.
              append it_ztimedata.
              MODIFY ztimedata from it_ztimedata.
           COMMIT WORK.
      ENDLOOP.
    in that coding if any correction pls correct me.
    Nikhil Narkhede

    Hi,
    You just try like this..Move date field into one Variable Year field into one variable and
    Time field into one variable..Then Pass this field to your screen fields.
    Rgs,
    Priya.

  • Upload Comma separated Text file.

    Hi all ,
    i would like to upload a comma seperated text file into an internal table using CL_GUI_FRONTEND_SERVICES=>GUI_UPLOAD but am not able to load correctly. Please let me know any aditional parameters should be pass to this FM.
    data **********
    060000012,1001,01052005,2.5,C123,0198345678912
    08000001,1001,01052005,5.0,C123,016545678912
    08000001,1002,02052005,2.5,C234,012345678912
    types  : Begin of ty_itab,
    pernr(10),
    ABS_ATT_TYPE(4),
    REC_CCTR(10),
    work_order(12),
    date(8),
    hours(8).
    types: end of ty_itab.
    data : itab type standard table of ty_itab with header line .
    CALL METHOD CL_GUI_FRONTEND_SERVICES=>GUI_UPLOAD
      EXPORTING
        FILENAME                = 'C:\LOAD.TXT'
         FILETYPE                = 'ASC'
        HAS_FIELD_SEPARATOR     = 'X'
       HEADER_LENGTH           = 0
       DAT_MODE                = SPACE
       CODEPAGE                = SPACE
       IGNORE_CERR             = ABAP_TRUE
       REPLACEMENT             = '#'
       READ_BY_LINE            = 'X'
    IMPORTING
       FILELENGTH              =
       HEADER                  =
      CHANGING
        DATA_TAB                = itab[]
      EXCEPTIONS
        FILE_OPEN_ERROR         = 1
        FILE_READ_ERROR         = 2
        NO_BATCH                = 3
        GUI_REFUSE_FILETRANSFER = 4
        INVALID_TYPE            = 5
        NO_AUTHORITY            = 6
        UNKNOWN_ERROR           = 7
        BAD_DATA_FORMAT         = 8
        HEADER_NOT_ALLOWED      = 9
        SEPARATOR_NOT_ALLOWED   = 10
        HEADER_TOO_LONG         = 11
        UNKNOWN_DP_ERROR        = 12
        ACCESS_DENIED           = 13
        DP_OUT_OF_MEMORY        = 14
        DISK_FULL               = 15
        DP_TIMEOUT              = 16
        NOT_SUPPORTED_BY_GUI    = 17
        ERROR_NO_GUI            = 18
        others                  = 19
    IF SY-SUBRC <> 0.
    MESSAGE ID SY-MSGID TYPE SY-MSGTY NUMBER SY-MSGNO
               WITH SY-MSGV1 SY-MSGV2 SY-MSGV3 SY-MSGV4.
    ENDIF.
    Thx
    Palani

    as you are telling that is comma delimited then save the file as comma delimited,
    this can be done uisng excel , open the text fiel with excel and save it as .csv file. and try ...
    or else i will tell another solution for it...
    Data: begin of itab occurs 0,
          rec(1000) type c,
          end of itab.
    Data: begin of itab2 occurs 0,
          matnr type mara-matnr
          werks type marc-werks
          quant(10) type c.
          end of itab2.
      file = 'C:test.txt'.
      call method cl_gui_frontend_services=>gui_upload
        exporting
          filename                = file
        filetype                = 'ASC'
         has_field_separator     = 'X'
        changing
          data_tab                = itab
      exceptions
        others                  = 17
      if sy-subrc <> 0.
    * MESSAGE ID SY-MSGID TYPE SY-MSGTY NUMBER SY-MSGNO
    *            WITH SY-MSGV1 SY-MSGV2 SY-MSGV3 SY-MSGV4.
      endif.
    Loop at itab.
    clear itab2.
    split itab-rec at ',' into itab2-matnr
                               itab2-werks
                               itab2-quant.
    append itab2.
    endloop.
    you can do using any of the one...

  • 8g touch updated 4.2.1, problems synching apps and music. gives multiple errors; device times out, internal device errors, network times out, duplicate file, too many files open. itunes diagnostics ok. what to do?

    have an 8g ipod touch which has had problems trying to play games and downloading update 4.2.1. went to the apple store who kindly downloaded the software update but now can't synch and generates multiple error messages; device times out, internal device error, network times out, duplicate file and too many files open etc etc. have latest i tunes download and diagnostic test ok. what can i do?

    Does the ext directory have the php_oci8.dll? In the original steps the PHP dir is renamed. In the given php.in the extension_dir looks like it has been updated correctly. Since PHP distributes php_oci8.dll by default I reckon there would be a very good chance that the problem was somewhere else. Since this is an old thread I don't think we'll get much value from speculation.
    -- cj

Maybe you are looking for