Unable to takeout spaces for type STPO-MENGE

Hi to all,
i struck up with one problem, v_amount is populating  value with spaces.
i am unable to get v_quantity value with out spaces.
i tried with SHIFT LEADING and Replace statements but it throws error message.
can any body help me regarding this
data : v_quantity type stpo-menge value '        12345'.

Hi ,
Try like below
data : v_quantity type stpo-menge value '        123.45'.
data : v_qua(20).
write: v_quantity to v_qua left-justified.
write v_qua.
Regards,
Himanshu

Similar Messages

  • After Effects error: Unable to allocate space for image buffer

    I am having some glitchy issues with After Effects, very inconsistent, but I am getting the error message "After Effects error: unable to Allocate Space for a 7500 x 4500 image buffer. You may be experiencing fragmentation. In the Memory and Multiprocessing Preference dialog box, trying increasing the RAM to leave for other applications, and selecting the Enable Disk Cache option in the Media & Disk Cache Panel".
    Pretty straightforward, but I have a brand new 8 core dual 2.25 Mac Pro with 16GB Ram. I have it set to allow other applications to use 7GB, have jockeyed that setting between 1GB and 5GB and still gotten the same error message. Disk Cache is set to 3GB, have pushed it all the up to 7GB and still gotten the same message. I am working in 1080p, but I got the error message with Open GL on, quarter resolution, and wasn't even trying to RAM preview. I also noticed that with a straight 1080p clip .mov animation codec clip, after effects will only RAM preview about 7 seconds. That is without any other layers or effects.  I do not have all of my source footage in one folder, I am going to try that next. I am running of a fast internal hard drive that is not they system drive.  It also seems like it happens after I have been using the comptuer for awhile. If I restart the problem will usually go away for a little while.
    I think that I may have a bigger system error, the computer locked up bad once after I had it for a few weeks and since that happened CS4 has been unstable in general, especially Photoshop and Bridge. My guess is bad RAM, but I wanted to make sure I wasn't missing something with my AE settings.  If any one has any input please let me know. Thanks.

    > I lowered the minimum RAM per core to 1GB, because the Adobe site recommends that as a base setting for andHD project
    When you say "the Adobe site recommends", what exactly are you referring to? If you're quoting recommendations written for After Effects on the Adobe website, you're very likely quoting me. And this is what I wrote in the Help document:
    "Memory & Multiprocessing preferences":
    "The amount of RAM required for each background process varies depending on your system configuration; at least 1 GB per process is recommended. Optimum performance is achieved with computer systems with at least 2 GB of installed RAM per processor core."
    This blog post gives essentially the same advice, but with more explicit suggestions.
    But, as I say at the bottom of that blog post, if you find that some other settings are working better for you, that's great. Every project and computer system are different. Do what works for you.
    > Also, if it is an
    dual quad core system, there should be 8 cores, but in the
    multi-processer preferences panel it lists CPU's as 16.
    The number of "virtual" processors can be double in a system that uses hyperthreading. After Effects doesn't actually treat these as separate processors in this context, though.

  • Removing leading spaces for type P variable

    Hi,
    I am declaring the 2 variables as type P. The below code was working fine, But our client, wants to remove the leading spaces for these variables,
    v_lower(16)           TYPE p DECIMALS 2,
    v_upper(16)           TYPE p DECIMALS 2.
    v_lower = i_plmk-toleranzun.
    v_upper = i_plmk-toleranzob.
    i_output-toleranzun  = v_lower.
    i_output-toleranzob  = v_upper.
    Can any one suggest how can we achieve this. Is there any function module for achieving this.
    Thanks In Advance,
    Regards,
    Ramana Trapatla.

    Hi,
    If your question was in the context of displaying data of type p without the leading space, then you can use left-justified addtion to your write statement.
    data: val(10) type p decimals 2.
    val = '0000000123.43'.
    write: val LEFT-JUSTIFIED.
    Regards,
    Vikranth

  • Space for 9708x15542

    Hi all,
    Hope everyone is O.K. well I am not, silly AE CS3 PRO. I have on my system Windows Vista 32bit intel Duo Core 6600 2.40ghz 4.0GB memory, and 300GIG of hard drive space free. I have created a basic 30sec comp using mainly 3D text and transistions. Nothing to strenouis. Anyhow when I render the comp a fault appears "UNABLE TO ALLOCATE SPACE FOR 9708X15542 IMAGE".
    Is it a memory issue/Cache?
    What do I need to do to fix this problem?
    Thank you all.
    josel

    That is a ginormous size -- close to IMAX resolution.  As Mylenium said, AE doesn't handle very large comps well. Other 32-bit composting apps handle them far better (like Shake & Nuke), so it's not necessarily a 32-bit memory issue. AE needs to generate very large image buffers if you have motion blur, 3D layers, or lights on very large layers.
    There's a useful guide from Jonas Hummelstrand at General Specialist that you might find helpful:
    http://generalspecialist.com/2006/11/avoiding-after-effects-error-could-not.asp
    Basically, you chop up your huge image into layers in Photoshop. Jonas was able to process 32,000 x 32,000 resolution layers using this method. Do you have any extremely large layers that could benefit from this method?

  • After Effects Error: Unable To Allocate Space...

    I have never seen this before...I am stumped.
    I am on After Effects 8.0.2 on a PC at work (Windows XP Pro 2002 SP2 / Intel Xeon 3.6 Ghz / 3.25 GB RAM) and I keep getting the following error when rendering:
    "After Effects Error: Unable to Allocate Space for 4290 x 4590 image buffer"
    My Comp is nowhere near that size (1280 x 720) nor is anything in it (PS document 1200 x 1500 / 720p video / solids are comp size). I have reset Cache sizes until I am blue in the face / have enabled Disk Cache / dumped preferences / cursed at the machine / nothing works.
    Any suggestions would be appreciated...thanks.
    Joey

    Just thinkin; out loud, so sorry if it's stuff you've already considered, Joey:
    Any scale settings expanding objects beyond the size of the comp? Text especially can blow out memory when scaled up large.
    Same goes for 3D objects close to the camera.
    As you're a Mac-o-file you've probably been saved in the past by the extra RAM accessible - SP will only give AE 2GB.

  • Error: Unable to fulfil request for 262144 bytes of memory space.

    Hi experts,
    When running a query and 30 minutes after that, I received an error message from BW:
    <b>Error Group
    RFC_ERROR_SYSTEM_FAILURE
    Message
    Unable to fulfil request for 262144 bytes of memory space.</b>
    Can someone help me with this?

    Hi Bob,
    Many thanks for your reply. I believe your observation is correct. I can run the report when I choose only one month instead of 17 months.
    However, do you think that something can be done on BW server to eliminate this problem? Is there not enough memory on BW server to fulfil this request?
    Thanks.

  • ORA-1653: unable to extend table - but enough space for datafile

    We encountered this problem in one of our database Oracle Database 10g Release 10.2.0.4.0
    We have all datafiles in all tablespaces specified with MAXSIZE and AUTOEXTEND ON. But last week database could not extend table size
    Wed Dec  8 18:25:04 2013
    ORA-1653: unable to extend table PCS.T0102 by 128 in                 tablespace PCS_DATA
    ORA-1653: unable to extend table PCS.T0102 by 8192 in                tablespace PCS_DATA
    Wed Dec  8 18:25:04 2013
    ORA-1653: unable to extend table PCS.T0102 by 128 in                 tablespace PCS_DATA
    ORA-1653: unable to extend table PCS.T0102 by 8192 in                tablespace PCS_DATA
    Wed Dec  8 18:25:04 2013
    ORA-1653: unable to extend table PCS.T0102 by 128 in                 tablespace PCS_DATA
    ORA-1653: unable to extend table PCS.T0102 by 8192 in                tablespace PCS_DATA
    Datafile was created as ... DATAFILE '/u01/oradata/PCSDB/PCS_DATA01.DBF' AUTOEXTEND ON  NEXT 50M MAXSIZE 31744M
    Datafile PCS_DATA01.DBF had only 1GB size. Maximum size is 31GB but database did not want to extend this datafile.
    We used temporary solution and we added new datafile to same tablespace. After that database and our application started to work correctly.
    There is enough free space for database datafiles.
    Do you have some ideas where could be our problem and what should we check?
    Thanks

    ShivendraNarainNirala wrote:
    Hi ,
    Here i am sharing one example.
    SQL> select owner,table_name,blocks,num_rows,avg_row_len,round(((blocks*8/1024)),2)||'MB' "TOTAL_SIZE",
      2   round((num_rows*avg_row_len/1024/1024),2)||'Mb' "ACTUAL_SIZE",
      3   round(((blocks*8/1024)-(num_rows*avg_row_len/1024/1024)),2) ||'MB' "FRAGMENTED_SPACE"
      4   from dba_tables where owner in('DWH_SCHEMA1','RM_SCHEMA_DDB','RM_SCHEMA') and round(((blocks*8/1024)-(num_rows*avg_row_len/1024/1024)),2) > 10 ORDER BY FRAGMENTED_SPACE;
    OWNER           TABLE_NAME                        BLOCKS   NUM_ROWS AVG_ROW_LEN TOTAL_SIZE           ACTUAL_SIZE          FRAGMENTED_SPACE
    DWH_SCHEMA1     FP_DATA_WLS                        14950     168507          25 116.8MB              4.02Mb               112.78MB
    SQL> select tablespace_name from dba_segments where segment_name='FP_DATA_WLS' and owner='DWH_SCHEMA1';
    TABLESPACE_NAME
    DWH_TX_DWH_DATA
    SELECT /* + RULE */  df.tablespace_name "Tablespace",
           df.bytes / (1024 * 1024) "Size (MB)",
           SUM(fs.bytes) / (1024 * 1024) "Free (MB)",
           Nvl(Round(SUM(fs.bytes) * 100 / df.bytes),1) "% Free",
           Round((df.bytes - SUM(fs.bytes)) * 100 / df.bytes) "% Used"
      FROM dba_free_space fs,
           (SELECT tablespace_name,SUM(bytes) bytes
              FROM dba_data_files
             GROUP BY tablespace_name) df
    WHERE fs.tablespace_name   = df.tablespace_name
    GROUP BY df.tablespace_name,df.bytes
    UNION ALL
    SELECT /* + RULE */ df.tablespace_name tspace,
           fs.bytes / (1024 * 1024),
           SUM(df.bytes_free) / (1024 * 1024),
           Nvl(Round((SUM(fs.bytes) - df.bytes_used) * 100 / fs.bytes), 1),
           Round((SUM(fs.bytes) - df.bytes_free) * 100 / fs.bytes)
      FROM dba_temp_files fs,
           (SELECT tablespace_name,bytes_free,bytes_used
              FROM v$temp_space_header
             GROUP BY tablespace_name,bytes_free,bytes_used) df
    WHERE fs.tablespace_name   = df.tablespace_name
    GROUP BY df.tablespace_name,fs.bytes,df.bytes_free,df.bytes_used
    ORDER BY 4 DESC;
    set lines 1000
    col FILE_NAME format a60
    SELECT SUBSTR (df.NAME, 1, 60) file_name, df.bytes / 1024 / 1024 allocated_mb,
    ((df.bytes / 1024 / 1024) - NVL (SUM (dfs.bytes) / 1024 / 1024, 0))
    used_mb,
    NVL (SUM (dfs.bytes) / 1024 / 1024, 0) free_space_mb
    FROM v$datafile df, dba_free_space dfs
    WHERE df.file# = dfs.file_id(+)
    GROUP BY dfs.file_id, df.NAME, df.file#, df.bytes
    ORDER BY file_name;
    Tablespace                      Size (MB)  Free (MB)     % Free     % Used
    DWH_TX_DWH_DATA                     11456       8298         72         28
    FILE_NAME                                                    ALLOCATED_MB    USED_MB FREE_SPACE_MB
    /data1/FPDIAV1B/dwh_tx_dwh_data1.dbf                                 1216       1216             0
    /data1/FPDIAV1B/dwh_tx_dwh_data2.dbf                                10240       1942          8298
    SQL> alter database datafile '/data1/FPDIAV1B/dwh_tx_dwh_data2.dbf' resize 5G;
    alter database datafile '/data1/FPDIAV1B/dwh_tx_dwh_data2.dbf' resize 5G
    ERROR at line 1:
    ORA-03297: file contains used data beyond requested RESIZE value
    Although , we did moved the tables into another TB , but it doesn't resolve the problem unless we take export and drop the tablespace aand again import it .We also used space adviser but in vain .
    As far as metrics and measurement is concerned , as per my experience its based on blocks which is sparse in nature related to HWM in the tablespace.
    when it comes to partitions , just to remove fragmentation by moving their partitions doesn't help  .
    Apart from that much has been written about it by Oracle Guru like you .
    warm regards
    Shivendra Narain Nirala
    how does free space differ from fragmented space?
    is all free space considered by you to be fragmented?
    "num_rows*avg_row_len" provides useful result only if statistics are current & accurate.

  • Query - unable to fulfill request for 65536 bytes of memory space

    Hi all,
             When i executing query it is running for 20 mins and showing this below error. Please guide me on this.
    Error:
    - Unable to fulfill request for 65536 bytes of memory space
    - No more storage space available for extending an internal table
    Note:
            Report design based on cube and it have around 60 laks records, aggrated and compression already available for this cube.
    Thanks & Regards,
    R. Saravanan

    Hi raj,
    Are you trying to execute the report in BEx analyzer?
    if data is huge try to reduce the volume of data with the help of filters for ex run the report for particular material, date ,cost center etc.
    If you are using BEx analyzer and with MS office -2003 version then Ms excell has a limitation of 65536 rows, this might be causing the error.
    Looks like your report is having more than 65536 records (selections for which your report was run).
    if you want all the records to be displayed try to use WAD. or use the MS excel 2007 or greater versions(not sure if it can hold all your 60 lack records).
    Regards
    KP
    Edited by: prashanthk on Feb 1, 2011 3:09 PM

  • Unable to fulfil request for 3665920 bytes of memory space.

    Hi Experts,
    I am facing the error " Unable to fulfil request for 3665920 bytes of memory space." when I try to call adobe form document from portal.The profile parameter ztta/max_memreq_MB has value of Min - 5 and Max - 2048.
    Please let me know your valuable suggestions to resolve the issue.
    Thanks in advance.
    Regards,
    Arun.

    Hello,
    I am facing the same issue as well.  However, this profile param has a max value of 2048 right?  The current value in our system is 2047.  It means that this can not be extended anymore.  If that is the case, is there any other way?
    Thank you.
    Mark

  • Unable to fulfil request for 3971508 bytes of memory space in SMQ2 -SYSFAIL

    Hi Experts,
    In Production Server PI SMQ2 transaction, I am getting SYSFAIL Status for a specific Queue Name <XBTO6___0000>. When I am double clicking it is showing "Unable to fulfil request for 3971508 bytes of memory space". I have tried to "Execute LUW" but the status is not changing at all. I have gone to the Message in SXMB_MONI, it is showing "Recorded for Outbound Processing".
    Please advice as it is continously taking more and more Queue Entries.
    Thanks,
    Nabendu.

    Hi Nabendu,
    May be this thread will help you...have a look..
    Re: Unable to fulfil request for 528700 bytes of stora ge space.
    Regds,
    Pinangshuk.

  • Unable to Extend TEMP space for CLOB

    Hi,
    I have a data extract process and I am writing the data to a CLOB variable. I'm getting the error "Unable to Extend TEMP space" for the larger extracts. I was wondering if writing the data to a CLOB column on a table and then regularly committing would be better than using a CLOB variable assuming time taken is not an issue.

    You do not need to add more temp files. This is not a problem of your temp tablespace. This is the problem of temp segments in your permanent tablespace. You need to add another datafile in EDWSTGDATA00 tablepsace. This happens when you are trying to create tables and indexes. Oracle first does the processing in temp segments(not temp tablespace) and at the end oracle converted those temp segments into permanent segments.
    Also, post the result of below query
    select file_name,sum(bytes/1024/1024)in_MB,autoextensible,sum(maxbytes/1024/1024)in_MB from dba_data_files where tablespace_name='STAGING_TEST' group by file_name,autoextensible order by FILE_NAME;

  • I rented a short film on iTunes but didn't have enough space for the HD version I paid for. I was unable to download it in SD for whatever reason and cannot seem to download my purchase on my desktop iTunes. Am I missing something?

    I rented a short film on iTunes but didn't have enough space for the HD version I paid for. I was unable to download it in OR convert it to SD for whatever reason and cannot seem to download my purchase on my desktop iTunes. Am I missing something?

    When you purchase an HD movie or show, in most cases the SD copy can be downloaded from your purchases area in iTunes. On the main iTunes Store page, go to Purchases under the Quick Links and then select Movies or TV shows as appropraite. Click the "Not On This Computer" button. Uncheck the "Download HD when available" box and you should then see the SD copies of the movies/shows and be able to download them.
    Note that with digital copies from a Blu-Ray or DVD, if it comes with an HD copy, you normally cannot get the SD copy as well and have to purchase it separately if you want it.
    Regards.

  • Unable to fulfil request for 528700 bytes of stora ge space.

    Hi all,
    I have a problem with messages getting stuck in XI, and
    if I check QRFC monitor I get a SYSFAIL with message
    "Unable to fulfil request for 528700 bytes of storage space."
    So nothing passes thru.
    Any idea where to start solving the error?
    Regards,
    Fredrik

    Hello,
    does anybody have a solution to this problem?
    We've also already enlarged the ztta / max _ memreq _ MB parameter to 2048. In ST22 i can see the SYSTEM_NO_ROLL dump. The message there is "Unable to fulfil request for 310373472 bytes of memory space". When the message is processed i can see in ST02 that the Max.use of Extended memory is growing to 1.290.240 (about 1.3GB).
    If somebody has a solution out there please let me know.
    Thanks
    regards
    Florian

  • Unable to release space from table

    Hi all,
    We are unable to release space from a table called TST03 even after deletion of records.
    Followings are the information.
    Database     : 9.2
    Table Name      : TST03
    Tablespace     : LOCALY MANAGED.
    Previously there were lots of rows.
    At Present only 9 No of Rows.
    Space allocated     : 41 GB
    PCT_INCREASE     : Null
    One of the column is of LONG RAW type.
    Since the table is in LMTS, we were expecting that the space allocated will be released automaticaly after deletion of records.
    Now, what are the option left with us to release 41GB of space ?
    A. Is there any effect of "Drop storage" option of 'Truncate table' command ?
    B. If yes , can i copy all the 9 rows to a new table, then use "Truncate table TST03 drop storage", check if space is released and then copy back all the 9 rows to this table.
    C. Do you have any other easy solution apart from export/import ?
    D. Checked all the relevant Note(646681,48400,10551), could not find an easy solution. I  want to avoid offline export/import option.
    Thanks .
    Naba J Neog

    > Hi all,
    Hi !
    > We are unable to release space from a table called
    > TST03 even after deletion of records.
    >
    > Followings are the information.
    > Database     : 9.2
    > Table Name      : TST03
    > Tablespace     : LOCALY MANAGED.
    > Previously there were lots of rows.
    > At Present only 9 No of Rows.
    > Space allocated     : 41 GB
    > PCT_INCREASE     : Null
    > One of the column is of LONG RAW type.
    >
    > Since the table is in LMTS, we were expecting that
    > the space allocated will be released automaticaly
    > after deletion of records.
    Sorry, but that is not what LMTS is for. Wrong assumption.
    MaxDB e.g. returns space immediately - Oracle does not.
    > Now, what are the option left with us to release 41GB
    > of space ?
    > A. Is there any effect of "Drop storage" option of
    > 'Truncate table' command ?
    Yes, the effect is, that after the truncate the table is empty and  only one extent is allocated - the rest is returned to freespace. That's the DEFAULT behaviour of TRUNCATE TABLE.
    > B. If yes , can i copy all the 9 rows to a new table,
    > then use "Truncate table TST03 drop storage", check
    > if space is released and then copy back all the 9
    > rows to this table.
    Nope - you would have to copy the LONG RAW columns as well and this cannot be done easily from sqlplus.
    > C. Do you have any other easy solution apart from
    > export/import ?
    Nope again - you'd have to use this offline reorganisation as long as you're not on Oracle 10g and the long raw fields got converted into LOBs. With 10g you might also use the SHRINK table command. But with Oracle 9i - sorry: exp and imp (will be pretty fast for 9 rows...)
    > D. Checked all the relevant Note(646681,48400,10551),
    > could not find an easy solution. I  want to avoid
    > offline export/import option.
    Sorry - no way to avoid it and still get the free space back.
    Anyhow you might want to take actions to prevent this situtation from reoccuring.
    These kinds of questions are covered in the notes
    <a href="http://service.sap.com/sap/support/notes/48400">#48400</a>
    <a href="http://service.sap.com/sap/support/notes/66290">#66290</a>
    >
    > Thanks .
    > Naba J Neog
    You're welcome.
    Lars

  • Need help with error on Lion Server (kernel: add_fsevent: unable to get path for vp ..)

    I have had some Xsan issues and one of my volumes kept failing. I did a complete hard reboot of the entire system after trying all the cvfsck options. Now the volume seems to be ok but I am now getting the following errors and can not figure out what they are. Any help to a new Admin would be greatly appreciated.
    5/7/12 1:59:13.000 PM kernel: add_fsevent: unabled to get a path for vp 0xb129000.  dropping the event.
    5/7/12 1:59:15.000 PM kernel: add_fsevent: unable to get path for vp 0xb12a4d0 (live.0.indexHead; ret 22; type 4)
    5/7/12 1:59:15.000 PM kernel: add_fsevent: unabled to get a path for vp 0xb12a4d0.  dropping the event.
    5/7/12 1:59:15.000 PM kernel: add_fsevent: unable to get path for vp 0xb129de0 (live.1.indexHead; ret 22; type 4)
    5/7/12 1:59:15.000 PM kernel: add_fsevent: unabled to get a path for vp 0xb129de0.  dropping the event.
    5/7/12 1:59:15.000 PM kernel: add_fsevent: unable to get path for vp 0xe3964d0 (live.2.indexHead; ret 22; type 4)
    5/7/12 1:59:15.000 PM kernel: add_fsevent: unabled to get a path for vp 0xe3964d0.  dropping the event.
    5/7/12 2:01:22.024 PM mdworker32: kCGErrorFailure: Set a breakpoint @ CGErrorBreakpoint() to catch errors as they are logged.
    5/7/12 2:01:22.525 PM com.apple.mdworker.pool.0: PSSniffer error: Invalid argument
    5/7/12 2:01:24.697 PM mdworker: FontImporter: Validation failed - "/Volumes/Matilda/Dexter/Randy project - Burke update/James E Burke Awards video revision/Past Video files/ETH742/tricyclestudios (After FX) Jeremy/Assets/FUTURA_4.TTF".
    5/7/12 2:01:24.697 PM mdworker: FontImporter: Validation Result - "(
        kATSFontTestSeverityFatalError
    Then I got a few more of the errors like the top lines.
    Please help!!
    Thanks,
    Kevin Rosenthal

    Turns out there were disk issues that I repaired by rebooting into Lion Recovery Mode (restart, hold Command-R during startup) and running Disk Utility. Disk Utility reported:
    Volume bitmap needs minor repair for orphaned blocks
    Invalid volume free block count
    It fixed both problems and the drive then verified clean.
    While I was there, I also repaired permissions, though I doubt that the permission oddities were the cause of the errors.
    After rebooting into normal Lion, two things happend:
    Time Machine quite happily went back to work and took a backup, no issue. Several hourlies have run since then, no problems.
    Spotlight did some re-indexing. This caused some hangs (spinning beachballs) at first, but it self-cleared. It definitely didn't last long enough to re-index the entire drive; I'm guessing that it needed to re-index some of the repaired sections.
    So far as I know, no files were corrupted -- but the drive is about 80% free space, so odds are that the orphaned blocks were not in a critical area.

Maybe you are looking for