R3load  Export issue

HI all
I got an error message while exporting. Any Suggestions.
Error 2 when calling FindMyChars
PROGNAME 2f 00 31 00 42 00 43 00 44 00 57 00 42 00 2f 00 |/.1.B.C.D.W.B./.|
PROGNAME 44 00 42 00 47 00 4c 00 5f 00 41 00 43 00 43 00 |D.B.G.L._.A.C.C.|
PROGNAME 54 00 5f 00 43 00 43 00 20 00 20 00 20 00 20 00 |T._.C.C. . . . .|
PROGNAME 20 00 20 00 20 00 20 00 20 00 20 00 20 00 20 00 | . . . . . . . .|
PROGNAME 20 00 20 00 20 00 20 00 20 00 20 00 20 00 20 00 | . . . . . . . .|
DYNPNUMBER 31 00 30 00 30 00 30 00                         |1.0.0.0.        |
R3STATE 41 00                                           |A.              |
(CNV) ERROR: code page conversion failed
             rc = 2
.--============.|                              RSCP - Error                            || Error from:             Codepage handling (RSCP)                     || code:  128  RSCPENOOBJ   No such object                              || Dummy module without real rscpmc[34]                                 || module: rscpmm   no:    2 line:    74                    T100: TS008 || TSL01: CPV  p3: Dummy-IPC   p4: rscpmc4_init                         |`--
'(DB) INFO: disconnected from DB
D:\usr\sap\DEV\SYS\exe\run/R3load.exe: job finished with 1 error(s)
D:\usr\sap\DEV\SYS\exe\run/R3load.exe: END OF LOG: 20060706000632

Hi
You need to check Basis patch level by SPAM. Update to latest.
Second, Check Kernel patch level by SM51. Update to latest.
Hope it may help.
Regards
Gurprinder

Similar Messages

  • R3load EXPORT tips for 1,5 TB MAXDB Database are welcome!

    Hello to ALL;
    this post is addressed to ALL MAXDB-GURU's !!
    I'v few questions to MAXDB Performance-Guy's for R3load Export on MAXDB!
    (remark to my person:
    I'm certified OS/DB migration consultant and have already done over 500 OS/DB migrations since 1995 successfully)
    BUT now I'm face with following setup for Export:
    HP- IA64 BL870c,  4 CPUu2019s Dual-Core + Hyperthreading activ (mean's 16 "CPU'S" shown via top or glance)
    64 GB RAM, HPUX V11.31;
    MAXDB  7.7.07
    ECC 5.0; Basis/Aba 640.22
    SAP-Kernel 640-Unicode Patch 381
    8 sapdatas are configured on unique 200GB LUN's and VG's in HP-UX; on storage-side the 8 220GB LUN's are located on (only) 9 300GB-Disks with Raid-5 Level and within HP-UX each VG/LVOL is mirrored via LVM to a second desasterdatacenter (200m distance)
    LOGFILES:
    2 x 4 GB LUN Raid-1 on Storage and via LVM  also mirrored to failover-center
    MAXDB-Datasize: 1600 GB Overall and within this 1350 GB used, TEMPUsage about 25 GB !
    MAXDB-Parameter-Settings
    MAXCPUu2019s 10  (4 x IA64 QuadCore+Hyperthreading shows 16 Cores within top and galcne
    Cache_SIZE= I/O Buffer Cache = 16 GB (2.000.000 Pages)
    Data-Cache-HitRatio:  99.61%
    Undo-Cache = 99,98%
    the following sapnote for MAXDB Peformance and Migrations are well known and already processes
    928037, 1077887, 1385089, 1327874, 954268, 1464560, 869267,
    My major problem is the export-runtime with over 4 days on the first dry-run (6 R3load's running on Linux-APPL-Server), and 46h on second runtime, 6 R3loads running on DB-Server IA64)
    the third trail-run was aborted by me after 48hours and 50% export-dump-space was written. In all 3 dry-runs, no more than approx 3.5GB DUMP-Space were written per hour!
    My first question to all MAXDB-Guru'S: How can I influence/optimize the TEMP - Area in MAXDB?? I didn't find any hint's in SDN or SAPNOTES or MaxDB Wiki or google....As fare as I know, this TEMP area "resides" within MAXDB-datafiles, thus it's seperated on my 48 datafiles, spreaded over 8 LUN/VG/disks. But I see LESS throughput on disk's and MAXDB-Kernel uses only ONE of ten's cpu-cores (approx 80% - 120% of 1000%).
    The throughput or cpu-usage doesn't change when I use 2 or 4 or 10 or 16 R3load' processes in parallel. The "result" is always the same: approx. 3,5 GB Export-Dump and 1 CPU MAX-DB Kernelprocess is used !
    so the BIG Question for me: WHERE is the bottleneck ?? (RAID-5 Disk-LUNS mirrored with LVM ???)
    on HP-UX scsi_queue_depth_length I'v increased default value from 8 to 32 to 64 to 128 --> no impact
    2nd question:
    I'v read  OS-Note 1327874 - FAQ: SAP MaxDB Read Ahead/Prefetch, and we are running MAXDB 7.7.07.19! (MaxDB 7.8 is not suppored via PAM for IA64 640-UC-Kernel) and as far as I understood, this parameter will no HELP me on EXPORT with primary-key-sequence ! is this correct?  THUS: which parameter HELPS for speeding up Export-Runtime?
    MAXCPU is set to 10, but ONLY 1 of them is used??
    so this post is for ALL MAXDB GURU'S!
    who will be intrested to contriubte on this "high-sophisticated" migration-project with 1.5TB MAXDB-Database-size and ONLY 24h Downtime !!
    all tips and hints are welcome and I will give us coninued updates to this running project until WE did a successfull migration job.
    PS: Import is not yet started, but should be done within vSphere 5 and SLES 11 SP1 on MAXDB 7.8 ....and of yours in parallel to export with migration monitor, BUT again a challenge: 200km distance from source to traget system !!!
    NICE PROJECT;
    best regards Alfred

    Hi Alfred,
    nice project ... just some simple questions:
    Did you open a message at SAP? Maybe you could buy some upgrade support, this could be usefull to get direct access to the SAP support...
    Which byte order do you use? I know Itanium could use both. But it should be different, otherwise you would use a backup for the migration.
    And the worst question, I do not even want to ask: What about your MAXCPUS parameter? Is it set to more than 1? This could be the problem why only one CPU is used.
    Best regards
    Christian

  • Bizzare export issue

    Here is a very weird export issue. Windows XP.
    Project is set at h264 1920x1080 mono audio 29.97 fps
    I have 30 shots that make up the sequence. Each shot is exactly 90 frames. I exported each sequence at h264 1920x1080 with one channel of audio. The first shot exports just fine and is 5 MB. The second shot exports at 10 MB. The third is 15 MB, the tenth shot is 50MB. But the picture is still 90 frames for each shot. BUT, when I export without audio each exported file is just 5 MB. If I look at the properties of each clip the duration of the video and audio is just 90 frames. If I open one of these clips in QuickTime Pro and re-export it as source then the files size goes back to 5 MB; same video, same audio yet with the correct file size.
    Anyone seen this before?
    Is there a setting that could be incorrect?
    -cablet

    Weird, man. Not sure what might cause that.

  • System copy using R3load ( Export / Import )

    Hi,
    We are testing System copy using R3load ( Export / Import ) using our production data.
    Environment : 46C / Oracle.
    while executing export, it takes more than 20 hours, for the data to get exported, we want to reduce the export time drastically. hence, we want to achieve it by scrutinizing the input parameters.
    during export, there is a parameter by name splitting the *.STR and *.EXT files for R3load.
    the default input for the above parameter is No, do not split STR and EXT files.
    My question 1 : If we input yes instead of No and split the *.STR and *.EXT files, will the export time get reduced ?
    My Question 2: If the answer is yes to Question 1, will the time reduced will be significant enough ? or how much percentage of time we can expect it to reduce, compare to a No Input.
    Best Regards
    L Raghunahth

    Hi
    The time of the export depends on the size of your database (and the size of your biggest tables) and your hardware capacity.
    My question 1 : If we input yes instead of No and split the *.STR and *.EXT files, will the export time get reduced ?
    In case you have a few very large tables, and you have multiple cpu's and a decent disk storage, then splitting might significantly reduce the export time.
    My Question 2: If the answer is yes to Question 1, will the time reduced will be significant enough ? or how much percentage of time we can expect it to reduce, compare to a No Input.
    As you did not tell us about your database size and hardware details there is no way to give you anything but very basic metrics. Did you specify a parallel degree at all, was your hardware idling for 20 hrs or fully loaded already?
    20 hrs for a 100gb database is very slow. It is reasonable (rather fast in my opinion)  for a 2tb database.
    Best regards, Michael

  • Crystal JRC PDF Export Issue

    We are experiencing a PDF export issue with Crystal JRC version 12.2.202.  The PDF report shows 1 of 3 pages but ONLY the first page is generated.  When this occurs, ALL other users that are exporting to PDF experience the same issue with any multiple page report.  I can't find any errors in the log that would indicate what the problem is.  To resolve the issue, I have to recycle the WebLogic application server that hosts  the JRC, which would indicate something gets "locked" in memory and gets cleared when the app server is bounced.
    Any ideas on how to resolve this ?

    Hi Ted,
    Thanks for the response!
    Though, the mentioned Sun JVM Bug explains the root-cause, the Bug status also state that it is not fixed yet.
    Also, we are reluctant to change the JVM Option to -Xint due to unknown impacts on JVM optimizations.
    Is there a patch released for CrystalReports to deal with this behavior? We are using CRJ runtime 11.5.7.620 and seeking for a patch on this version.
    Thanks.

  • PP CS5.5 FLV To MP4 Export Issues / Problems

    I just upgraded from Production Premium CS5 to CS5.5. I do a fair amount of converting/exporting of imported FLV files (downloaded from internet news sites) to H.264/AAC w/ MP4 wrappers using Premiere Pro (PP) and Adobe Media Encoder (AME). My equipment is a 1+ yr. old HP Win 7 Pro w/ an Intel i970 Quadcore, 12 Gb RAM and MSI nVidia GTX 470 GPU. I have never experienced any queue or export issues w/ CS5 (at least that I noticed), but I am experiencing noticeably degraded queue export and direct export problems with these files in CS5.5.
    Description: (1) The queue export times for the exact same imported FLV clip(s) using the exact same sequence settings and H.264/MP4 export settings are noticeably slower in CS5.5 (sometimes 2x, depending on the export settings, e.g. using maximum render quality) than in CS5; (2) My MSI Afterburner GPU monitoring software indicates 5-10% GPU usage during queue/export in CS5, while indicating 0% GPU usage (flatline) in CS5.5; and (3) When I attempt to direct export these files using CS5.5 I get the following error message: "The source and output audio channels are not compatible or a conversion does not exist."
    I suspect that it may have something to do with "conforming" anomalies when the FLV file is imported into PP, but I would put that squarely in the category of rank speculation on my part.
    Has anyone experienced the same or similar problems or have any insight into what may be the problem(s)? Of course, thanks for any help/insight you can provide.
    Jim

    What do you mean by a better routine?
    A better compression routine than H.264.
    Is the quality better than H.264?
    Yes.
    Does this use an MOV wrapper?
    Yes.
    Also, would it import correctly into PPro?
    Yes.
    we would import files into PPro exported out of FCP and would get a red pixelated glitch during preview and export. Does this codec solve this problem as well if it is an MOV wrapper?
    I don't know anything about FCP.
    Sorry for all the questions, just want to get things right before diving into a new codec.
    It's harmless, and free.
    Download it and run some tests.

  • Simple flash export issue

    hello
    I'm trying to export a flash video using Adobe media encoder from Premier Pro, but when I do I get this error " there are no clips to build".
    I have another file that I am able to export and it does not have this problem.
    They are Windows Media files and play in Premier fine. I am using Premier Pro CS3.
    Lastly, I'm not a video expert but I really need to get these up on the web for work, there is a lot of pressure and I need to get it done.
    Regards,
    Tom

    fixed the exporting issue.

  • R3Load (Export _System Copy )

    Dear Experts,
    System Copy ( Source Version )
    Environment : R/3 4.6C  / Oracle 8.1.7 / Solaris 8
    using R3Load ( Export / Import ) procedure.
    We have a Problem.
    As the Database is around 430 GB.
    We need approx 140 GB for the Export Image File.
    But what we have is not more than 50 GB on each Hard disk. (We have many hard disks but free space not more than 50 GB in any Disk.)
    So we are thinking of Splitting the STR Files and do the export.
    But is this possible?.
    Actually I want to specify multiple locations for the Export Files.
    or can you suggest me other alternative..
    Thanking you in Anticipation.
    Best Regards
    L Raghunahth

    hi
    Thanks for Markus and Michael for your suggestions.
    I think I should opt for NFS thing as it is a Production System.
    Next, our system is R/3 4.6C ( 46D 64 Bit Kernel, Current Patch Level 1748)
    for just executing R3Load Export, is this patch level is ok, as I could not find any prerequisites regarding the kernel requirements.
    Our customer is little hesitant for applying the Latest Kernel Patches as it is the PRD System.
    I wish to know the minimum requirement of Kernel Patch for successful execution of R3Load Export only.
    Could you please let me know where i can find for this prerequisite.
    Thanks in Advance.
    Best Regards
    L Raghunahth

  • Exporting ISSUES! Help

    I am a relatively new user to lightroom.  I haven't had any exporting issues until today.  I am attempting to export a photo/photos to a specific folder with a custom name.  When I do this it exports the WHOLE library and NOT the edited photo.  What am I doing wrong??

    Lightroom usually exports all images that are highlighted. If you only want to export one image then make sure it is the only image that is highlighted.

  • Exporting issues after 4.2 upgrade

    Has anyone else experienced exporting issues after installing the 4.2 upgrade? It used to be a breeze but now I keep getting the message that a destination folder has not been specified even though I've done that. Any feedback would be greatly appreciated! Thank you.

    Are you using an Export preset, or setting the Export up manually each time? 

  • How to run and R3load export remotely?

    Hi,
    I need to know  how to do an R3load export from an application server instead of a Central instance.
    Does anyone did this? Do you have documentation about it?
    Thanks all in advance
    Diego,

    This should work if you log directly into the application server instance. Remember that an application server is just a set of work processes, each of which has a connection to the database. So if you log into the app server as sidadm, then you should be able to run R3load from the application server.
    Hope that helps.
    J. Haynes

  • PREPARE before R3load Export/Import?

    We are planning an upgrade from SAP 4.7 to ECC 6.0, where the system will also be moved onto new hardware with new OS and Oracle versions.  Our procedure is to perform a Homogeneous System Copy of our current SAP 4.7 system, using R3load to complete an Export, then using R3load to Import the existing system onto the new hardware.
    Our current system setup, where the Export will be completed is:
    SAP 4.7 (Domain Installation)
    Windows 2000
    Oracle 9.2.0.5.
    Our new system, SAP 4.7 (Local Installation), will be Imported using the R3load procedure onto our new hardware with Windows 2003 and Oracle 10.2.0.2.  Once the SAP 4.7 copy has been completed on the new hardware, we will begin the SAP upgrade to ECC 6.0.
    Based on our procedure, am I still able to effectively complete the PREPARE on our current system before the R3load Export is ran?  Will this PREPARE still be effective once the system has been Imported (copied) onto different versions of OS and Oracle?  Or, will I need to re-run the PREPARE once it has been copied onto the new os and Oracle versions?  Any experience would be very helpful in making this decision.
    Thanks in advance.
    Tina

    It might be best to wait until after the system copy for PREPARE.
    1) If this machine is your transport domain controller, there will be files in the transport directory on the source system that you will need to copy to the new system.
    2) If you are migrating from 32-bit to 64-bit Windows, PREPARE will load the 32-bit kernel tools on the source system but you will need the 64-bit kernel on the new system.
    3) You would need to copy /usr/sap/put contents from the source system.
    In addition, running PREPARE after the system copy would provide a validation of the system copy procedures.  I have been through this process and PREPARE has always been run after the system copy to the new hardware.

  • Reporting Export Issue from Taleo

    Can anyone help?  We have a work around for the Exporting Issues into Excel 2010 regarding figures and dates only reading as text, however, a new issue has arisen.   In one field we have Aggregated Agencies assigned to any particular Job Code.   In the legacy reports, every agency name exported into one cell in Excel, now they export into individual rows, this causes a report to be too long and there is no way to amalgamate them into one cell.    Has anyone found a workaround for this?

    is there someway that it was reformatted by the Iomega drive being daisy chained to it?
    No. That is not possible -anyway, reformatting erases everything on a drive.
    You would have noticed that first.
    You can download the [NTFS-3G driver|http://macntfs-3g.blogspot.com> which will let you read, write and even reformat the clients drive as NTFS on the Mac. All reasonably modern Windows PCs work with this disc format and you won't run into the file size limitations associated with FAT formatted drives.

  • Using 128bit key Blowfish and any export issues

    Hi,
    I want to use blowfish (symmetric key algorithm) in my product and which will be shipped outside US. Reading Sun's documentation I am bit confused and I am not sure if the jce package is going to be available outside the US.
    Has anyone used the default available symmetric key algorithms in the JCE package (javax.crypto.*) and faced any export control issues ?
    Please let me know.
    Thanks
    RT
    Here is a sample code
    String ks = "123456789123456";
    kb = ks.getBytes();
    SecretKeySpec key = new SecretKeySpec(kb,"blowfish");
    Cipher cipher = Cipher.getInstance("blowfish");
    // Initialisation of cipher object for Encryption.
    cipher.init(Cipher.ENCRYPT_MODE, key);

    1. the issue is rather an import issue for some countries, than an export issue for US as far as you use jre>=1.4 .
    you can read the
    'Downloading the "Unlimited Strength" Jurisdiction Policy Files'
    section of
    http://java.sun.com/products/jce/javase.html
    2. The restricted (defualt) shipped version already covers Blowfish if you mind that.
    3. I use Blowfish with restricted (default) version of JCE in applications deployed in europe and have not heard of customer complaining sofar: they are supposed to use webstart or jre1.5.

  • R3load export of  table REPOSRC with lob col - error ora-1555 and ora-22924

    Hello,
    i have tried to export data from our production system for system copy and then upgrade test. while i export the R3load job has reported error in table REPOSRC, which has lob column DATA. i have apsted below the conversation in which i have requested SAP to help and they said it comes under consulting support. this problem is in 2 rows of the table.
    but i would like to know if i delete these 2 rows and then copy from our development system to production system at oracle level, will there be any problem with upgrade or operation of these prorgams and will it have any license complications if i do it.
    Regards
    Ramakrishna Reddy
    __________________________ SAP SUPPORT COnveration_____________________________________________________
    Hello,
    we have are performing Data Export for System copy of our Production
    system, during the export, R3load Job gave error as
    R3LOAD Log----
    Compiled Aug 16 2008 04:47:59
    /sapmnt/DB1/exe/R3load -datacodepage 1100 -
    e /dataexport/syscopy/SAPSSEXC.cmd -l /dataexport/syscopy/SAPSSEXC.log -stop_on_error
    (DB) INFO: connected to DB
    (DB) INFO: DbSlControl(DBSL_CMD_NLS_CHARACTERSET_GET): WE8DEC
    (DB) INFO: Export without hintfile
    (NT) Error: TPRI_PAR: normal NameTab from 20090828184449 younger than
    alternate NameTab from 20030211191957!
    (SYSLOG) INFO: k CQF :
    TPRI_PAR&20030211191957&20090828184449& rscpgdio 47
    (CNV) WARNING: conversion from 8600 to 1100 not possible
    (GSI) INFO: dbname = "DB120050205010209
    (GSI) INFO: vname = "ORACLE "
    (GSI) INFO: hostname
    = "dbttsap "
    (GSI) INFO: sysname = "AIX"
    (GSI) INFO: nodename = "dbttsap"
    (GSI) INFO: release = "2"
    (GSI) INFO: version = "5"
    (GSI) INFO: machine = "00C8793E4C00"
    (GSI) INFO: instno = "0020111547"
    (DBC) Info: No commits during lob export
    DbSl Trace: OCI-call 'OCILobRead' failed: rc = 1555
    DbSl Trace: ORA-1555 occurred when reading from a LOB
    (EXP) ERROR: DbSlLobGetPiece failed
    rc = 99, table "REPOSRC"
    (SQL error 1555)
    error message returned by DbSl:
    ORA-01555: snapshot too old: rollback segment number with name "" too
    small
    ORA-22924: snapshot too old
    (DB) INFO: disconnected from DB
    /sapmnt/DB1/exe/R3load: job finished with 1 error(s)
    /sapmnt/DB1/exe/R3load: END OF LOG: 20100816104734
    END of R3LOAD Log----
    then as per the note 500340, i have chnaged the pctversion of table
    REPOSRC of lob column DATA to 30, but i get the error still,
    i have added more space to PSAPUNDO and PSAPTEMP also, still the same
    error.
    the i have run the export as
    exp SAPDB1/sap file=REPOSRC.dmp log=REPOSRC.log tables=REPOSRC
    exp log----
    dbttsap:oradb1 5> exp SAPDB1/sap file=REPOSRC.dmp log=REPOSRC.log
    tables=REPOSRC
    Export: Release 9.2.0.8.0 - Production on Mon Aug 16 13:40:27 2010
    Copyright (c) 1982, 2002, Oracle Corporation. All rights reserved.
    Connected to: Oracle9i Enterprise Edition Release 9.2.0.8.0 - 64bit
    Production
    With the Partitioning option
    JServer Release 9.2.0.8.0 - Production
    Export done in WE8DEC character set and UTF8 NCHAR character set
    About to export specified tables via Conventional Path ...
    . . exporting table REPOSRC
    EXP-00056: ORACLE error 1555 encountered
    ORA-01555: snapshot too old: rollback segment number with name "" too
    small
    ORA-22924: snapshot too old
    Export terminated successfully with warnings.
    SQL> select table_name, segment_name, cache, nvl(to_char
    (pctversion),'NULL') pctversion, nvl(to_char(retention),'NULL')
    retention from dba_lobs where
    table_name = 'REPOSRC';
    TABLE_NAME | SEGMENT_NAME |CACHE | PCTVERSION | RETENTION
    REPOSRC SYS_LOB0000014507C00034$$ NO 30 21600
    please help to solve this problem.
    Regards
    Ramakrishna Reddy
    Dear customer,
    Thank you very much for contacting us at SAP global support.
    Regarding your issue would you please attach your ORACLE alert log and
    trace file to this message?
    Thanks and regards.
    Hello,
    Thanks for helping,
    i attached the alert log file. i have gone through is, but i could
    not find the corresponding Ora-01555 for table REPOSRC.
    Regards
    Ramakrishna Reddy
    +66 85835-4272
    Dear customer,
    I have found some previous issues with the similar symptom as your
    system. I think this symptom is described in note 983230.
    As you can see this symptom is mainly caused by ORACLE bug 5212539 and
    it should be fixed at 9.2.0.8 which is just your version. But although
    5212539 is implemented, only the occurrence of new corruptions will be
    avoided, the already existing ones will stay in the system regardless of the patch.
    The reason why metalink 452341.1 was created is bug 5212539, since this
    is the most common software caused lob corruption in recent times.
    Basically any system that was running without a patch for bug 5212539 at some time in the past could be potentially affected by the problem.
    In order to be sure about bug 5212539 can you please verify whether the
    affected lob really is a NOCACHE lob? You can do this as described in
    mentioned note #983230. If yes, then there are basically only two
    options left:
    -> You apply a backup to the system that does not contain these
    corruptions.
    -> In case a good backup is not available, it would be possible to
    rebuild the table including the lob segment with possible data loss . Since this is beyond the scope of support, this would have to be
    done via remote consulting.
    Any further question, please contact us freely.
    Thanks and regards.
    Hello,
    Thanks for the Help and support,
    i have gone through  the note 983230 and metalink 452341.1.
    and i have ran the script and found that there are 2 rows corrupted in
    the table REPOSRC. these rows belong to Standard SAP programs
    MABADRFENTRIES & SAPFGARC.
    and to reconfirm i have tried to display them in our development system
    and production system. the development systems shows the src code in
    Se38 but in production system it goes to short dump DBIF_REPO_SQL_ERROR.
    so is it possible to delete these 2 rows and update ourselves from our
    development system at oracle level. will it have any impact on SAP
    operation or upgrade in future.
    Regards
    Ramakrishna Reddy

    Hello, we have solved the problem.
    To help someone with the same error, what we have done is:
    1.- wait until all the processes has finished and the export is stopped.
    2.- startup SAP
    3.- SE14 and look up the tables. Crete the tables in the database.
    4.- stop SAP
    5.- Retry the export (if you did all the steps with sapinst running but the dialogue window in the screen) or begin the sapinst again with the option: "continue with the old options".
    Regards to all.

Maybe you are looking for

  • Default value at ME21N

    Hi i want account assignment 'K' and item catagory 'D' get default if user selects perticular document type while creating Purchase order. if PO creation date is greater than a perticular date it should be applicable. screen field account assignment

  • Opening new Pages files in Old Pages

    Seem to have a problem opening pages files created in the new version. Any way of sorting this. Can I update Pages? Don't want to buy new version. Cheers in advance

  • Load a director level into flash // is it possible?

    Hi, I am planning a complex system where eventually users will need to do immersion into 3rd dimension and also play lots of flash quizz. I know director can load flash which makes it obvious choice for development but coding in lingo (a non POO lang

  • Differences between Synchronous and Asynchronous updates

    can any one tell me the differences between these updates

  • New prebuilt Database App Developer VM - posted 16 th Nov2011

    Note: as of 16th November 2011 there is a new image that has been updated with Apex4.1, new Hands-on Labs content, includes links to streamed interviews with Development and Engineering. You can download it here http://www.oracle.com/technetwork/data