Conversion of Reservation to PR

Hi SAP Gurus,
Let me know is there any facility to convert reservation (Whose material are non-stock) directly to Purchase requision? Instead of creation of PR manually by user?
Waiting for your valuable reply.
VB

VB,
Reservation is a kind of requirement whilest purchase requisistion is a procurement proposal. They are two, totally different things.
E.g. header material is A, children part is "B" that is "A"'s BOM (let's say 1 pc "A" is produced 1 pc "B"). You create a PrdOrd of 10 pcs for material "A" > this will generate an order reservation of 10 pcs for material "B" > you run MRP > MRP generates a PurReq of 10 pcs for material "B" (depends on master data settings).
materual "B"
OrdRes 10 pcs -
PurReq 10 pcs +
So, procurement proposal is generated via MRP run for the requirement, but you cannot "convert" a requirement into a PR or PO, etc.
Regards,
Csaba

Similar Messages

  • Automatic conversion of reserved html character

    A client is wanting to know if there is any special way to handle
              conversion of reserved html characters.
              Specifically, he has heard that some web servers can assign other
              values to represent these characters that are converted upon
              presentation of the web page.
              Has anyone heard of this feature?
              Thanks,
              Mark
              

    Hi,
    Yawmark gave you an answer in this thread:
    http://forum.java.sun.com/thread.jspa?threadID=573268&messageID=4078415#4078415
    Kaj

  • File Content Conversion: Remove reserved characters from Content

    Hi PI Gurus,
    we're facing a little problem with FCC in File Receiver Adapter.  Sometimes the defined fieldSeparator ("|") comes within the field content.
    e.g.
    <?xml version="1.0" encoding="UTF-8"?>
    <ns0:Positions xmlns:ns0="http://test.com">
    <Position>
    <Value1>001</Value1>
    <Value2>Test|Content</Value2>
    </Position>
    </ns0:Positions>
    So, my first thought was to replace all "|" with a litte java mapping beind the actual mapping.
    But, is there any possibility to achieve this via configuration of FCC (something like fieldSeperatorSubstitution)?
    Help is as always highly appreciated!
    Cheers,
    Matthias Kral

    Hello Hareenkumar,
    yes you're totally right. I ran in several OutOfMemoryError while testing the code above. The errors started with messages over 15 MB.
    Here's the code, that is productive now. It replaces all Pipes "|" by Slashes "|" and really works fine and performant.
    public void execute(InputStream in, OutputStream out)
              throws StreamTransformationException {
              try {
                   int totalCount = 0;
              byte[] buf = new byte[40960];
              int count = 0;
                   while (count >= 0) {
                        count = in.read(buf);
                        totalCount += count;
                        if (count > 0) {
                             for (int i=0; i<count; i++){
                                  if (buf<i>==124){
                                       buf<i> = 47;
                             out.write(buf, 0, count);
              } catch (IOException e) {
                   throw new StreamTransformationException(e.getMessage());
    Thank you all again!!!
    Cheers Matthias

  • Case conversion for SQL and PL/SQL reserved key words

    Hi!
    Do we have available in Developer automatic conversion of the reserved words to upper cases?
    For example in vim, it is possible to customize such way that after typing a reserved word : 'create' as am example and I type next a space it automatically converts it to upper case: 'CREATE'
    Thanks,
    Radu

    You can do it afterwards with Format (CTRL-F7), but that will rearrange everything in your worksheet.
    Also using code completion, the reserved words are always lower case. This is very disturbing, as I have to fix the case afterwards and lose more time than I initially won using the code completion.
    I recon the case should be converted automatically upon typing, using the settings from the formatter.
    In case development doesn't agree right away, I've added a feature request: http://htmldb.oracle.com/pls/otn/f?p=42626:39:1041303283166847::NO::P39_ID:10621
    Please go vote to add weight to see it implemented.
    Thanks,
    K.

  • Supplier conversion into R12 giving error at site level(AP_INVALID_TCA_ERR)

    Hi All,
    I am running "Supplier Sites Open Interface Import" for Supplier Site conversion but standard program is failing with the following error msg: "AP_INVALID_TCA_ERROR".
    Can any one please help me how I can fix this issue. Any help will be Gold for me.
    Best Regards,
    Shaik.

    Hi Hussein,
    OS: Linux.
    DataBase: Oracle 11g
    Application is : Oracle R12.2
    Below is the coucurrent request Log.
    Payables: Version : 12.0.0
    Copyright (c) 1979, 1999, Oracle Corporation. All rights reserved.
    APXSSIMP module: Supplier Sites Open Interface Import
    Current system time is 09-NOV-2010 02:48:40
    +-----------------------------
    | Starting concurrent program execution...
    +-----------------------------
    Arguments
    P_WHAT_TO_IMPORT='ALL'
    P_COMMIT_SIZE='50000'
    P_PRINT_EXCEPTIONS='N'
    P_DEBUG_SWITCH='Y'
    P_TRACE_SWITCH='Y'
    APPLLCSP Environment Variable set to :
    Current NLS_LANG and NLS_NUMERIC_CHARACTERS Environment Variables are :
    American_America.AL32UTF8
    Enter Password:
    MSG-00001: After SRWINIT
    MSG-00002: After Get_Company_Name
    MSG-00003: After Get_NLS_Strings
    MSG-00004: After Importing Suppliers
    MSG-00005: After Get_Header_Information
    MSG-00020: After SRWEXIT
    Report Builder: Release 10.1.2.3.0 - Production on Tue Nov 9 02:48:43 2010
    Copyright (c) 1982, 2005, Oracle. All rights reserved.
    Executing request completion options...
    Output file size:
    2369
    ------------- 1) PRINT   -------------
    Printing output file.
    Request ID : 1446582      
    Number of copies : 0      
    Printer : noprint
    Finished executing request completion options.
    Concurrent request completed successfully
    Current system time is 09-NOV-2010 02:48:54
    Best Regards,
    Shaik

  • XML data from BLOB to CLOB - character set conversion

    Hi All,
    I'm trying to solve a problem with a character set conversion in PL/SQL in the following scenario:
    1. source is an XML as a BLOB variable.
    2. target is an XML as a CLOB variable.
    3. the problem I have is the following:
    - database character set is set to UTF-8
    - XML character set could be anything (UTF-8, ISO 8859-1, ISO 8859-2, ASCII, ...)
    - I need to write a procedure which converts the source BLOB content into the target CLOB taking into account the XML encoding and converts it into the DB default character set (UTF8).
    I've been able to implement a simple conversion function. However, this function expects static XML encoding ISO-8859-1. The main part of the function looks as follows:
    buffer := UTL_RAW.cast_to_varchar2(
    UTL_RAW.convert(
    DBMS_LOB.SUBSTR(source_blob_variable, 16000, pos)
    , 'American_America.UTF8'
    , 'American_America.we8iso8859p1')
    Does anyone have an idea how to rewrite the code to handle "any" XML encoding in the source BLOB file? In other words, is there a function in Oracle which converts XML character set names into Oracle character set values (ISO-8859-1 to we8iso8859p1, UTF-8 to UTF8, ...)?
    Thanks a lot for any help.
    Julius

    I want to pass a BLOB to some "createXML" procedure and get a proper XMLType in UTF8 character set, properly converted from whatever character set is the input in.As per documentation the generated XML has always the encoding set at the client side depending on NLS_LANG (default UTF-8), regardless of the input encoding, so I don't see a need to parse the PI of the XML:
    C:\>echo %NLS_LANG%
    %NLS_LANG%
    C:\>sqlplus
    SQL*Plus: Release 11.1.0.6.0 - Production on Wed Apr 30 08:54:12 2008
    Copyright (c) 1982, 2007, Oracle.  All rights reserved.
    Connected to:
    Oracle Database 11g Enterprise Edition Release 11.1.0.6.0 - Production
    With the Partitioning, OLAP, Data Mining and Real Application Testing options
    SQL> var cur refcursor
    SQL>
    SQL> declare
      2     b   blob := utl_raw.cast_to_raw ('<a>myxml</a>');
      3  begin
      4     open :cur for select xmlroot (xmltype (utl_raw.cast_to_varchar2 (b))) xml from dual;
      5  end;
      6  /
    PL/SQL procedure successfully completed.
    SQL>
    SQL> print cur
    XML
    <?xml version="1.0" encoding="UTF-8"?><a>myxml</a>
    SQL> exit
    Disconnected from Oracle Database 11g Enterprise Edition Release 11.1.0.6.0 - Production
    With the Partitioning, OLAP, Data Mining and Real Application Testing options
    C:\>set NLS_LANG=GERMAN_GERMANY.WE8ISO8859P1
    C:\>sqlplus
    SQL*Plus: Release 11.1.0.6.0 - Production on Mi Apr 30 08:55:02 2008
    Copyright (c) 1982, 2007, Oracle.  All rights reserved.
    SQL> var cur refcursor
    SQL>
    SQL> declare
      2     b   blob := utl_raw.cast_to_raw ('<a>myxml</a>');
      3  begin
      4     open :cur for select xmlroot (xmltype (utl_raw.cast_to_varchar2 (b))) xml from dual;
      5  end;
      6  /
    PL/SQL-Prozedur erfolgreich abgeschlossen.
    SQL>
    SQL> print cur
    XML
    <?xml version="1.0" encoding="ISO-8859-1"?>
    <a>myxml</a>

  • Conversion of 1.2 JATO project (non-Forte IDE) to S1AF (JATO) 2.0

    JATO Team,
    First of all, thank you very much for the Studio integration. It
    looks very promising in terms of development time consumption,
    deployment to S1AS7, etc., etc. It is impossible to observe all
    advantages for so short period of time.
    The SOAF (JATO) 2.0 is installed on the top of the Sun ONE Studio 4,
    update 1 (EE) with JDK 1.4 along with the Sun One Application Server
    7 (W2000). Everything looks okay, at least all basic tasks listed in
    the "Getting Started" manual (project/view/model creation, basic db
    connection, running of a test application with the usage of the
    Studio's default Tomcat container, etc.) work proper.
    During the installation of S1AF module on the top of the Sun ONE
    Studio 4, update 1 (EE) I've got an invitation to convert the
    existing project to the new environment. As I realized the only JATO
    project integrated/created with Forte IDE is applicable for this auto-
    conversion (please correct me, if I am wrong. It could solve a lot of
    my problems).
    Since we are using JBuilder5 IDE and our JATO 1.2 project is
    integrated with this IDE I manually re-created project in the Studio
    with its file structure, adjusted the project web.xml file, etc. This
    new project looks like a proper one (Studio recognize all methods,
    fields, bean patterns, e.g.) except at least one "small" thing. All
    java files (project viewBeans, Models, custom viewBeanHelpers a.k.a.
    pure java) came up with the same wizard image:java class. Stutio with
    the S1AF module reserves the special images (and appropriate studio
    properties, of course!) for View and Model. Namely this allows to see
    the following structure for a ViewBean, for example, in the Studio:
    Modules
    ProjectName.ModuleName
    FooViewBean
    JavaSource
    JSPs
    Non-Visual Components
    View Components
    On the other hand, I could add either new View or Model in my
    manually converted project and add any View Component or bind the
    Model fields, for example. Also, the ProjectModuleServlet has been
    converted proper. I tried to convert each View/Model class inside the
    Studio to its proper wizard image and failed. Addition of a View
    Components (with an appropriate code fragments) via the Studio or
    auto-binding of model fields is an essential part of 2.0 and should
    drastically speed up the development process as it is designed.
    Questions:
    1. What I missed in my manual conversion of 1.2 JATO project to the
    SOAF (JATO) 2.0 realm? What should I correct in my JATO 1.2
    compatible classes (Views and Models) to be recognizable by Studio
    wizard (JATO 2.0)?
    To be more specific. Some deltas between JATO 1.2 and JATO 2.0 are as
    follows (related to a ViewBean):
    ====++++++++++++++======
    JATO 1.2 viewBean extension upon creation:
    public class FooViewBeanViewBean extends ViewBeanBase
    implements ViewBean
    ========================
    JATO 2.0 (S1AF) version: viewBean extension upon creation:
    public class FooViewBean extends BasicBeanBase
    ====++++++++++++++======
    JATO 1.2 viewBean createChild(...) method for one static looks
    like:
    protected View createChild(String name)
    if (name.equals(CHILD_STATICTEXT1))
    StaticTextField child = new StaticTextField
    (this,
    getDefaultModel(),
    CHILD_STATICTEXT1,
    CHILD_STATICTEXT1,
    CHILD_STATICTEXT1_RESET_VALUE,
    null);
    return child;
    ========================
    JATO 2.0 (S1AF) version: viewBean createChild(...) is renamed (as
    a least) to createChildReserved(...):
    protected View createChildReserved(String name) {
    if (name.equals(CHILD_STATICTEXT1)) {
    com.iplanet.jato.view.BasicDisplayField child =
    new com.iplanet.jato.view.BasicDisplayField(this,
    STATICTEXT1);
    return child;
    ====++++++++++++++======
    JATO 1.2 viewBean registerChildren() method for the basic field
    types looks like:
    private void registerChildren() {
    registerChild(CHILD_STATICTEXT1, StaticTextField.class);
    registerChild(CHILD_TEXTFIELD1, TextField.class);
    registerChild(CHILD_BUTTON1, Button.class);
    ========================
    JATO 2.0 (S1AF) version: viewBean registerChildren() method:
    private void registerChildren() {
    registerChild(CHILD_STATICTEXT1,
    com.iplanet.jato.view.BasicDisplayField.class);
    registerChild(CHILD_TEXTFIELD1,
    com.iplanet.jato.view.BasicDisplayField.class);
    registerChild(CHILD_BUTTON1,
    com.iplanet.jato.view.BasicCommandField.class);
    Is it correct to say that all existing custom Views/Models
    (compatible with JATO 1.2) should be converted to their JATO 2.0
    variants to be visible by the Studio?
    On the other hand, fast overview of 2.0 API shows that the JATO 1.2
    is a sub-set of the 2.0 (BasicViewBean extends ViewBeanBase, for
    example). What exactly (only deprecated methods?) should be adjusted
    in the project code (1.2), if necessary, to be visible by Studio as a
    View or Model component?
    2. Where it is possible to find the list of deprecated methods (from
    JATO 1.2 to JATO 2.0 versions). It is difficult sometimes to compare
    two versions of API docs (1.2 and 2.0) or compare logs between
    versions. Also, JATO 2.0 is significantly larger than 1.2 version. If
    the later obviously inherits the ND conversion stage (via the
    previous versions), the former obviously has additions incoming from
    the integration with the Studio (a.k.a. Forte 4.0).
    3. What is the current/future Sun/JATO Team policy with regards to
    JATO source code access (version 2.0, at least)? The reason of this
    question is as follows: in order to display dates formatted on the
    screens we adjusted a couple of JATO 1.2 core classes, for example.
    The only minimal, absolutely necessary changes in JATO 1.2 was made,
    but anyway...
    Sorry for so long multiple question. As Todd wrote in his
    announcement mail: "We think you will be very pleased with the new
    product...". This is exciting moment, I believe, for JATO Team as
    well as for any team involved into the conversion of the full size
    application/product from ND to J2EE realm with the JATO as
    an application framework (1.0, 1.1.x, 1.2.X, and 2.0 finally). The
    last step is left in this spiral process to enjoy the ND_Studio
    attractive features in the open source environment.
    Thank you very much in advance.
    Vladimir Ivanov

    I'll get some file templates ASAP and provide them to the group.
    As for source code, I'm not sure what the policy is. JATO 1.2 is very
    close to what JATO 2.0 is, so that should suffice for now.
    The community will be informed when we know more about source availibility.
    craig
    vivanov4114 wrote:
    Craig,
    Thank you for the answer. To be honest with you I tried to do exactly
    the same: I created the small test project and made an attempt to
    adjust the existing *.xml files to the new ones. I got some results,
    otherwise I couldn't even see my original 1.2 project in the Studio.
    Since I am a newcomer in the Studio, I definitely missed something in
    my adjustments. I'll try to observe my changes with the fresh head.
    On the other hand, I am afraid that my samples are very pure. If you
    could post or send me examples of jatoapp.xml and web.xml files (say
    for any of you test project) or excerptions from them with the
    appropriate patterns (ViewBean and Model peers, at least), I would be
    very appreciated. My mail is: vivanov@b...
    In the worst-case scenario I see the workaround: to re-create the
    project for one of our releases with the Forte 4.1 IDE and auto-
    convert it into the JATO 2.0/Studio world using the Studio
    facilities. We must get the current project fully
    integrated/converted with/to the Studio (at certain point) because,
    first, we expect massive coding with the GUI components involved soon
    and, secondly, we have around thousand classes related to JATO only
    (and a lot of extras).
    Coming back to the question #3 from this post. Now we have full
    access to the version 1.2, not only to the JATO 1.2 jar file(I
    explained some reasons below). Would we expect the same Sun/JATO Team
    policy with JATO 2.0?
    Thank you again,
    Vlad
    --- In SunONE-JATO@y..., "cvconover" <craig.conover@s...> wrote:
    It doesn't appear that anyone has responded to this so I am goingto
    give you the short answer:
    The reason you don't see your ViewBeans and Models showing up with
    there "wizard created" icons is because of just that. They weren't
    created via the wizards. The wizards create xml formatted filesthat
    contain metadata of how/what to generate for the ViewBeans and
    Models.
    ViewBeans will have a .viewbean fiel, Models will have a .sqlmodel
    file (for SQL Models), etc. The content of these files describesthe
    code that needs to be generated in the corresponding Java classfile.
    So LoginViewBean.java will have a peer LoginViewBean.viewbean file.
    The code that is generated is place in protected code blocks thatcan
    not be modified in the Studio and should not be modified outsidethe
    Studio because the code will likely be regenerated and custom mods
    inside the protected blocks will be lost.
    Now you can make a JATO 1.2 app appear in the Studio just like aJATO
    2.0 app by adding a jatoapp.xml file with the proper content and
    adjusting your web.xml properly, but it's much more work to getyour
    v1.2 ViewBeans and Models to show up like wizard created versions.
    Furthermore, even more work to get the display fields to show up
    visually.
    The good news is that v1.2 ViewBeans will work with newly wizard
    created ViewBeans. And if you do go through the trouble of making
    your ViewBeans Studio visible like your wizard created ViewBeans,
    adding new display fields visually will work along side yourmanually
    created fields.
    Try creating a new JATO app using the Studio wizards and then go to
    the Filesystems tab and look for the jatoapp.xml file and theweb.xml
    file.
    I am looking for an email that I have that explains how to do a
    partial, manual upgrade.
    Also, you will get rid of your JATO 1.2 jar and replace with theJATO
    2.0 jar in your lib dir.
    Hope this will suffice for now.
    craig
    --- In SunONE-JATO@y..., "vivanov4114" <vivanov@b...> wrote:
    JATO Team,
    First of all, thank you very much for the Studio integration. It
    looks very promising in terms of development time consumption,
    deployment to S1AS7, etc., etc. It is impossible to observe all
    advantages for so short period of time.
    The SOAF (JATO) 2.0 is installed on the top of the Sun ONE Studio4,
    update 1 (EE) with JDK 1.4 along with the Sun One ApplicationServer
    7 (W2000). Everything looks okay, at least all basic tasks listedin
    the "Getting Started" manual (project/view/model creation, basic
    db
    connection, running of a test application with the usage of the
    Studio's default Tomcat container, etc.) work proper.
    During the installation of S1AF module on the top of the Sun ONE
    Studio 4, update 1 (EE) I've got an invitation to convert the
    existing project to the new environment. As I realized the onlyJATO
    project integrated/created with Forte IDE is applicable for thisauto-
    conversion (please correct me, if I am wrong. It could solve a
    lot
    of
    my problems).
    Since we are using JBuilder5 IDE and our JATO 1.2 project is
    integrated with this IDE I manually re-created project in theStudio
    with its file structure, adjusted the project web.xml file, etc.This
    new project looks like a proper one (Studio recognize all
    methods,
    fields, bean patterns, e.g.) except at least one "small" thing.All
    java files (project viewBeans, Models, custom viewBeanHelpersa.k.a.
    pure java) came up with the same wizard image:java class. Stutiowith
    the S1AF module reserves the special images (and appropriate
    studio
    properties, of course!) for View and Model. Namely this allows tosee
    the following structure for a ViewBean, for example, in the
    Studio:
    Modules
    ProjectName.ModuleName
    FooViewBean
    JavaSource
    JSPs
    Non-Visual Components
    View Components
    On the other hand, I could add either new View or Model in my
    manually converted project and add any View Component or bind the
    Model fields, for example. Also, the ProjectModuleServlet hasbeen
    converted proper. I tried to convert each View/Model class insidethe
    Studio to its proper wizard image and failed. Addition of a View
    Components (with an appropriate code fragments) via the Studio or
    auto-binding of model fields is an essential part of 2.0 and
    should
    drastically speed up the development process as it is designed.
    Questions:
    1. What I missed in my manual conversion of 1.2 JATO project tothe
    SOAF (JATO) 2.0 realm? What should I correct in my JATO 1.2
    compatible classes (Views and Models) to be recognizable byStudio
    wizard (JATO 2.0)?
    To be more specific. Some deltas between JATO 1.2 and JATO 2.0are
    as
    follows (related to a ViewBean):
    ====++++++++++++++======
    JATO 1.2 viewBean extension upon creation:
    public class FooViewBeanViewBean extends ViewBeanBase
    implements ViewBean
    ========================
    JATO 2.0 (S1AF) version: viewBean extension upon creation:
    public class FooViewBean extends BasicBeanBase
    ====++++++++++++++======
    JATO 1.2 viewBean createChild(...) method for one static looks
    like:
    protected View createChild(String name)
    if (name.equals(CHILD_STATICTEXT1))
    StaticTextField child = new StaticTextField
    (this,
    getDefaultModel(),
    CHILD_STATICTEXT1,
    CHILD_STATICTEXT1,
    CHILD_STATICTEXT1_RESET_VALUE,
    null);
    return child;
    ========================
    JATO 2.0 (S1AF) version: viewBean createChild(...) is renamed(as
    a least) to createChildReserved(...):
    protected View createChildReserved(String name) {
    if (name.equals(CHILD_STATICTEXT1)) {
    com.iplanet.jato.view.BasicDisplayField child =
    new com.iplanet.jato.view.BasicDisplayField(this,
    STATICTEXT1);
    return child;
    ====++++++++++++++======
    JATO 1.2 viewBean registerChildren() method for the basic field
    types looks like:
    private void registerChildren() {
    registerChild(CHILD_STATICTEXT1,
    StaticTextField.class);
    registerChild(CHILD_TEXTFIELD1, TextField.class);
    registerChild(CHILD_BUTTON1, Button.class);
    ========================
    JATO 2.0 (S1AF) version: viewBean registerChildren() method:
    private void registerChildren() {
    registerChild(CHILD_STATICTEXT1,
    com.iplanet.jato.view.BasicDisplayField.class);
    registerChild(CHILD_TEXTFIELD1,
    com.iplanet.jato.view.BasicDisplayField.class);
    registerChild(CHILD_BUTTON1,
    com.iplanet.jato.view.BasicCommandField.class);
    Is it correct to say that all existing custom Views/Models
    (compatible with JATO 1.2) should be converted to their JATO 2.0
    variants to be visible by the Studio?
    On the other hand, fast overview of 2.0 API shows that the JATO1.2
    is a sub-set of the 2.0 (BasicViewBean extends ViewBeanBase, for
    example). What exactly (only deprecated methods?) should beadjusted
    in the project code (1.2), if necessary, to be visible by Studio
    as
    a
    View or Model component?
    2. Where it is possible to find the list of deprecated methods(from
    JATO 1.2 to JATO 2.0 versions). It is difficult sometimes tocompare
    two versions of API docs (1.2 and 2.0) or compare logs between
    versions. Also, JATO 2.0 is significantly larger than 1.2
    version.
    If
    the later obviously inherits the ND conversion stage (via the
    previous versions), the former obviously has additions incomingfrom
    the integration with the Studio (a.k.a. Forte 4.0).
    3. What is the current/future Sun/JATO Team policy with regards
    to
    JATO source code access (version 2.0, at least)? The reason ofthis
    question is as follows: in order to display dates formatted onthe
    screens we adjusted a couple of JATO 1.2 core classes, forexample.
    The only minimal, absolutely necessary changes in JATO 1.2 wasmade,
    but anyway...
    Sorry for so long multiple question. As Todd wrote in his
    announcement mail: "We think you will be very pleased with the
    new
    product...". This is exciting moment, I believe, for JATO Team as
    well as for any team involved into the conversion of the fullsize
    application/product from ND to J2EE realm with the JATO as
    an application framework (1.0, 1.1.x, 1.2.X, and 2.0 finally).The
    last step is left in this spiral process to enjoy the ND_Studio
    attractive features in the open source environment.
    Thank you very much in advance.
    Vladimir IvanovTo download the latest version of JATO, please visit:
    http://www.sun.com/software/download/developer/5102.html
    For more information about JATO, please visit:
    http://developer.iplanet.com/tech/appserver/framework/index.jsp
    [Non-text portions of this message have been removed]

  • Best Practice for Conversion Workflow

    Hello,
    I'm converting video files from our "home grown" virtual media reserve to iTunes U. Some of the files are in RM format, some are already compressed .mov's (not H.264) and some I have the original DV files for.
    Anyone out there have a best practice for converting these file types for posting to iTunes U? I have Final Cut Studio (Compressor), QT Pro and Squeeze available to me.
    Any experience you have with this would be helpful.
    Thanks,
    Jeana

    For converting old files to a podcast compatible video and based on the machine you have, consider elgato turbo.264. It is a fairly priced "co-processor" for video conversion. It is comprised of an application and a small USB device with a encoder chip in it. In my experience, it is the fastest way to create podcast video files. The amount of time that you will save will pay for the device quickly (about $100). Plus it does batch conversion of any video that your system currently plays through QuickTime. it has all the necessary presets and you can create your own. It has a few minor limitations such as not supporting (at this moment) enhanced podcasting features such as chapter markers and closed-captions but since you have old files for conversion, that won't matter.
    For creating new content, the workflow varies a lot. Since you mention MP3s, I guess you are also interested in audio files. I would stick with GarageBand, especially if you are a beginner plus it supports enhanced podcasts.
    In any case the most important goal is to have the simplest and fastest way to go from recording to publication. The less editing the better. To attain that, the best methods will require the largest investments. For example, for video production the best way is to produce the content live so when you finish recording it is only a matter of encoding and publishing. that will require the use of a video switcher that can ingest at least one video camera and a computer output to properly capture presentation material. That's the minimum. there are several devices that can do this for you. Some are disguised PCs and some connect to a PC for tapeless recording. You can check the Tricaster, which I like but wish it was a Mac and not a Windows Xp PC. Other routes may include video mixers from manufacturers like Edirol, Pansonic or Sony connected to a VTR or directly to a Mac for direct-ti-disc capture. I f you look at some of the content available in iTunes U, you will see what I explain here. This workflow requires preparation and sufficient live support but you will have your material ready for delivery almost immediately after the recorded event. No editing required. Finally, the most intensive workflow is to record everything separately and edit it later, which is extremely time consuming.

  • PM Work Order Conversion Advice

    Experts,
    We are currently implementing ECC 6.08 and have a requirement to convert open and historical work orders during the implementation phase.  While we have already selected a BAPI to load the data into the target system, we are now struggling to understand what financial impacts we must watch out for, and how to reconcile related documents (purchase orders, reservations, etc.).
    Has any one completed a conversion of Open PM Work Orders (transaction IW31) in the past u2013 either manually or via an automated conversion?
    If so u2013 how did you manage the complexities of related documents, reconciliation of financial transactions, etc?
    Any advice, old documentation, or planning materials that you may have from a previous project would be fantastic!
    Thanks in advance!

    Hi Ranganathan,
    You can use Tcode IW61 to create Historical Orders. The historical orders has no goods movements & settlement. You can directly enter the actual cost during historical order creation.Those orders can be used for cost analysis.
    Regards,
    senthil m

  • Error in Conversion of Planned Orders to AFS production Orders

    Dear Mam/Sir,
    We are working in AFS Scenario.There are following observations during
    AFS BOM/Planned order conversion/Material availabilty as follows:
    1) When Item category 'T'(Text Item) is used ,with 'witout fixed qty. indicator'in AFS BOM,then during the MRP Run BOM gets exploded & dependent requirements are created for AFS /non-AFS Components. But, when Bill of Material Tab is accessesd in the Planned Order system throws error'REQUIRED PARAMETER MISSING WHEN CALLIN UP MODULE MARC_SINGLE_READ'.When this planned order is converted to Production order, Order reservation is created only for NON-AFS(std. component).
    2) When a Planned order(with all BOM components item  category 'L') is converted ,then system gives error in material availability check : Message no.AG022'An error occurred during Remote Function call on the ATP Server'.
    3) When for a AFS material BOM components with item category'L' is used an order is created.during material availabilty check system shows missing part. The moment  a 'text' item is added in BOM the for the same order system commits all components.Though components are missing.
    Kindly advise.
    regards,
    Stuti

    Hi,
    Good to know that your problem is resolved.
    But the error you mentioned appears may be because of configuration problems.
    You should check what is account assignment category mentioned in the corresponding requirement class for the strategy group you have used in the materials master.
    Also you should check the settlement profile that you have used for your order types.
    The account assignment category and settlement profile maintained, must be set properly so that allowed transactions are executed.
    Also, there is problem with the message itself which appears for many reasons. More particularly when you are in MTO / ETO scenario or using WBS to create orders.This make it difficult to check the correct cause of the problem!!
    Pavan

  • Conversions between character sets when using exp and imp utilities

    I use EE8ISO8859P2 character set on my server,
    when exporting database with NLS_LANG not set
    then conversion should be done between
    EE8ISO8859P2 and US7ASCII charsets, so some
    characters not present in US7ASCII should not be
    successfully converted.
    But when I import such a dump, all characters not
    present in US7ASCII charset are imported to the database.
    I thought that some characters should be lost when
    doing such a conversions, can someone tell me why is it not so?

    Not exactly. If the import is done with the same DB character set, then no matter how it has been exported. Conversion (corruption) may happen if the destination DB has a different character set. See this example :
    [ora102 work db102]$ echo $NLS_LANG
    AMERICAN_AMERICA.WE8ISO8859P15
    [ora102 work db102]$ sqlplus test/test
    SQL*Plus: Release 10.2.0.1.0 - Production on Tue Jul 25 14:47:01 2006
    Copyright (c) 1982, 2005, Oracle.  All rights reserved.
    Connected to:
    Oracle Database 10g Enterprise Edition Release 10.2.0.1.0 - Production
    With the Partitioning, OLAP and Data Mining options
    TEST@db102 SQL> create table test(col1 varchar2(1));
    Table created.
    TEST@db102 SQL> insert into test values(chr(166));
    1 row created.
    TEST@db102 SQL> select * from test;
    C
    ¦
    TEST@db102 SQL> exit
    Disconnected from Oracle Database 10g Enterprise Edition Release 10.2.0.1.0 - Production
    With the Partitioning, OLAP and Data Mining options
    [ora102 work db102]$ export NLS_LANG=AMERICAN_AMERICA.EE8ISO8859P2
    [ora102 work db102]$ sqlplus test/test
    SQL*Plus: Release 10.2.0.1.0 - Production on Tue Jul 25 14:47:55 2006
    Copyright (c) 1982, 2005, Oracle.  All rights reserved.
    Connected to:
    Oracle Database 10g Enterprise Edition Release 10.2.0.1.0 - Production
    With the Partitioning, OLAP and Data Mining options
    TEST@db102 SQL> select col1, dump(col1) from test;
    C
    DUMP(COL1)
    ©
    Typ=1 Len=1: 166
    TEST@db102 SQL> exit
    Disconnected from Oracle Database 10g Enterprise Edition Release 10.2.0.1.0 - Production
    With the Partitioning, OLAP and Data Mining options
    [ora102 work db102]$ echo $NLS_LANG
    AMERICAN_AMERICA.EE8ISO8859P2
    [ora102 work db102]$ exp test/test file=test.dmp tables=test
    Export: Release 10.2.0.1.0 - Production on Tue Jul 25 14:48:47 2006
    Copyright (c) 1982, 2005, Oracle.  All rights reserved.
    Connected to: Oracle Database 10g Enterprise Edition Release 10.2.0.1.0 - Production
    With the Partitioning, OLAP and Data Mining options
    Export done in EE8ISO8859P2 character set and AL16UTF16 NCHAR character set
    server uses WE8ISO8859P15 character set (possible charset conversion)
    About to export specified tables via Conventional Path ...
    . . exporting table                           TEST          1 rows exported
    Export terminated successfully without warnings.
    [ora102 work db102]$ sqlplus test/test
    SQL*Plus: Release 10.2.0.1.0 - Production on Tue Jul 25 14:48:56 2006
    Copyright (c) 1982, 2005, Oracle.  All rights reserved.
    Connected to:
    Oracle Database 10g Enterprise Edition Release 10.2.0.1.0 - Production
    With the Partitioning, OLAP and Data Mining options
    TEST@db102 SQL> drop table test purge;
    Table dropped.
    TEST@db102 SQL> exit
    Disconnected from Oracle Database 10g Enterprise Edition Release 10.2.0.1.0 - Production
    With the Partitioning, OLAP and Data Mining options
    [ora102 work db102]$ imp test/test file=test.dmp
    Import: Release 10.2.0.1.0 - Production on Tue Jul 25 14:49:15 2006
    Copyright (c) 1982, 2005, Oracle.  All rights reserved.
    Connected to: Oracle Database 10g Enterprise Edition Release 10.2.0.1.0 - Production
    With the Partitioning, OLAP and Data Mining options
    Export file created by EXPORT:V10.02.01 via conventional path
    import done in EE8ISO8859P2 character set and AL16UTF16 NCHAR character set
    import server uses WE8ISO8859P15 character set (possible charset conversion)
    . importing TEST's objects into TEST
    . importing TEST's objects into TEST
    . . importing table                         "TEST"          1 rows imported
    Import terminated successfully without warnings.
    [ora102 work db102]$ export NLS_LANG=AMERICAN_AMERICA.WE8ISO8859P15
    [ora102 work db102]$ sqlplus test/test
    SQL*Plus: Release 10.2.0.1.0 - Production on Tue Jul 25 14:49:34 2006
    Copyright (c) 1982, 2005, Oracle.  All rights reserved.
    Connected to:
    Oracle Database 10g Enterprise Edition Release 10.2.0.1.0 - Production
    With the Partitioning, OLAP and Data Mining options
    TEST@db102 SQL> select col1, dump(col1) from test;
    C
    DUMP(COL1)
    ¦
    Typ=1 Len=1: 166
    TEST@db102 SQL>

  • Conversion of csv in flat file to oracle table format

    hi friends
    I have written a C code for conversion of flat file data to Comma seperated values
    Now i have to put this .csv file into oracle tables.
    Please suggest me a solution to this so that i can put this values into oracle tables
    Regards,
    Mahesh

    You can use Sql*Loader (sqlldr) :
    TEST@db102 SQL> create table test (a varchar2(30), b varchar2(30), c varchar2(30));
    Table created.
    TEST@db102 SQL> exit
    Disconnected from Oracle Database 10g Enterprise Edition Release 10.2.0.1.0 - Production
    With the Partitioning, OLAP and Data Mining options
    $ cat test.csv
    aaa,bbb,ccc
    ddd,eee,fff
    111,222,333
    $ cat test.ctl
    load data
    infile 'test.csv'
    replace into table test
    fields terminated by ','
    (a char, b char, c char)
    $ sqlldr test/test control=test.ctl
    SQL*Loader: Release 10.2.0.1.0 - Production on Wed Mar 8 15:42:42 2006
    Copyright (c) 1982, 2005, Oracle.  All rights reserved.
    Commit point reached - logical record count 3
    $ sqlplus test/test
    SQL*Plus: Release 10.2.0.1.0 - Production on Wed Mar 8 15:42:48 2006
    Copyright (c) 1982, 2005, Oracle.  All rights reserved.
    Connected to:
    Oracle Database 10g Enterprise Edition Release 10.2.0.1.0 - Production
    With the Partitioning, OLAP and Data Mining options
    TEST@db102 SQL> select * from test;
    A                    B                    C
    aaa                  bbb                  ccc
    ddd                  eee                  fff
    111                  222                  333
    TEST@db102 SQL>                                                                                                             But probably there's no need to convert the flat file to csv...

  • Conversion failed when converting the varchar value 'undefined' to data typ

    Conversion failed when converting the varchar value 'undefined' to data type int.
    hi, i installed oracle insbridge following the instruction in the manual. in rate manager, when i tried to create a new "Normal rating" or "Underwriting", im getting the following exception
    Server Error in '/RM' Application.
    Conversion failed when converting the varchar value 'undefined' to data type int.
    Description: An unhandled exception occurred during the execution of the current web request. Please review the stack trace for more information about the error and where it originated in the code.
    Exception Details: System.Data.SqlClient.SqlException: Conversion failed when converting the varchar value 'undefined' to data type int.
    Source Error:
    An unhandled exception was generated during the execution of the current web request. Information regarding the origin and location of the exception can be identified using the exception stack trace below.
    Stack Trace:
    [SqlException (0x80131904): Conversion failed when converting the varchar value 'undefined' to data type int.]
    System.Data.SqlClient.SqlConnection.OnError(SqlException exception, Boolean breakConnection) +1948826
    System.Data.SqlClient.SqlInternalConnection.OnError(SqlException exception, Boolean breakConnection) +4844747
    System.Data.SqlClient.TdsParser.ThrowExceptionAndWarning(TdsParserStateObject stateObj) +194
    System.Data.SqlClient.TdsParser.Run(RunBehavior runBehavior, SqlCommand cmdHandler, SqlDataReader dataStream, BulkCopySimpleResultSet bulkCopyHandler, TdsParserStateObject stateObj) +2392
    System.Data.SqlClient.SqlDataReader.HasMoreRows() +157
    System.Data.SqlClient.SqlDataReader.ReadInternal(Boolean setTimeout) +197
    System.Data.SqlClient.SqlDataReader.Read() +9
    System.Data.SqlClient.SqlCommand.CompleteExecuteScalar(SqlDataReader ds, Boolean returnSqlValue) +50
    System.Data.SqlClient.SqlCommand.ExecuteScalar() +150
    Insbridge.Net.Fwk.DAO.DataAccess.ScalarQuery(String connectionString, String command, Transaction transType, Object[] procParams) +110
    [Exception: Cannot Execute SQL Command: Conversion failed when converting the varchar value 'undefined' to data type int.]
    Insbridge.Net.Fwk.DAO.DataAccess.ScalarQuery(String connectionString, String command, Transaction transType, Object[] procParams) +265
    Insbridge.Net.Fwk.DAO.SqlProcessor.ExecuteScalarQueryProc(String subscriber, String datastore, String identifier, String command, Transaction transType, Object[] procParams) +101
    Insbridge.Net.Fwk.DAO.SqlProcessor.ExecuteScalarQuery(String subscriber, String identifier, String command) +22
    Insbridge.Net.RM.IBRM.ExeScalar(String cmd, Object[] paramsList) +99
    Insbridge.Net.RM.Components.Algorithms.AlgEdit.Page_Load(Object sender, EventArgs e) +663
    System.Web.Util.CalliHelper.EventArgFunctionCaller(IntPtr fp, Object o, Object t, EventArgs e) +14
    System.Web.Util.CalliEventHandlerDelegateProxy.Callback(Object sender, EventArgs e) +35
    System.Web.UI.Control.OnLoad(EventArgs e) +99
    System.Web.UI.Control.LoadRecursive() +50
    System.Web.UI.Page.ProcessRequestMain(Boolean includeStagesBeforeAsyncPoint, Boolean includeStagesAfterAsyncPoint) +627
    my insbridge versions are as follows
    IBRU 4.0.0 Copyright ©2010, Oracle. All rights reserved. - Version Listing
    RMBUILD.DLL 4.0.0.0 (x86)
    SRLOAD.DLL 3.13.0 (x86)
    IBRM v04.00.0.17
    IB_CLIENT v04.00.0.00
    RM.DLL 4.00.0 (x86)
    IBFA 3.0.2
    OS: Windows Server 2003
    DB: Sql Server 2008
    Browser: IE8
    how do i solve this, please help

    This is an error due to conversion failed from character string to int datatype. Infact, the table column contains "NO" value which you are trying to convert to Int/SUM which is illegal. Without your code and table structure, its difficult to pinpoint your
    actual issue. But check, all columns for value "NO". 

  • Available bandwindth and 'max-reserved bandwidth'

    Is the max-reserved bandwidth only important when working with Qos classes and the bandwidth statement? Is the default 75% available bandwidth only used then?
    In other words if I have a 100MB link with a service policy applied for Voice, Call-Control and video. After that I notice the available bandwidth on thie 100MB link is 61280 kilobits/sec.
    If I put in a 'max-reserved bandwidth 95' would I reclaim another 20MB of bandwidth for the class-default? Would leaving 5% on the 100MB link for routing and other stuff be acceptable?
    Here is the config and show commands:
    class-map match-any Call-Control
    match ip dscp cs3
    match ip dscp af31
    class-map match-any Video
    match ip dscp af41
    class-map match-any Voice
    match ip dscp ef
    policy-map QOS_classes_to_ACN
    class Voice
    priority 10000
    class Call-Control
    bandwidth 500
    class Video
    bandwidth 3220
    class class-default
    fair-queue
    random-detect
    interface FastEthernet6/0
    description 100MB Link to ACN
    ip address xxx.xxx.xxx.xxx xxx.xxx.xxx.xxx
    ip route-cache flow
    no ip mroute-cache
    load-interval 30
    duplex full
    speed 100
    service-policy output QOS_classes_to_ACN
    ROC-RT7206-QMOE#sh int f6/0
    FastEthernet6/0 is up, line protocol is up
    Hardware is i82543 (Livengood), address is 00b0.4a28.3ca8 (bia 00b0.4a28.3ca8)
    Description: 100MB Link to ACN
    Internet address is xxx.xxx.xxx.xxx/xx
    MTU 1500 bytes, BW 100000 Kbit, DLY 100 usec,
    reliability 255/255, txload 183/255, rxload 21/255
    Encapsulation ARPA, loopback not set
    Keepalive set (10 sec)
    Full-duplex, 100Mb/s, 100BaseTX/FX
    ARP type: ARPA, ARP Timeout 04:00:00
    Last input 00:00:03, output 00:00:00, output hang never
    Last clearing of "show interface" counters 01:13:30
    Input queue: 0/75/0/0 (size/max/drops/flushes); Total output drops: 5211742
    Queueing strategy: Class-based queueing
    Output queue: 70/1000/64/5211742 (size/max total/threshold/drops)
    Conversations 2/35/256 (active/max active/max total)
    Reserved Conversations 2/2 (allocated/max allocated)
    Available Bandwidth 61280 kilobits/sec <--- Available bandwidth
    30 second input rate 8615000 bits/sec, 6860 packets/sec
    30 second output rate 71788000 bits/sec, 7484 packets/sec
    31692173 packets input, 4263195179 bytes
    Received 1204 broadcasts, 0 runts, 0 giants, 0 throttles
    0 input errors, 0 CRC, 0 frame, 0 overrun, 0 ignored
    0 watchdog
    0 input packets with dribble condition detected
    34536300 packets output, 2513155446 bytes, 0 underruns
    0 output errors, 0 collisions, 0 interface resets
    0 babbles, 0 late collision, 0 deferred
    0 lost carrier, 0 no carrier
    0 output buffer failures, 0 output buffers swapped out

    Here is the output of show policy-pam int:
    ROC-RT7206-QMOE#sh policy-map int f6/0
    FastEthernet6/0
    Service-policy output: QOS_classes_to_ACN
    Class-map: Voice (match-any)
    3417571 packets, 934178998 bytes
    30 second offered rate 1722000 bps, drop rate 0 bps
    Match: ip dscp ef (46)
    3417571 packets, 934178998 bytes
    30 second rate 1722000 bps
    Queueing
    Strict Priority
    Output Queue: Conversation 264
    Bandwidth 10000 (kbps) Burst 250000 (Bytes)
    (pkts matched/bytes matched) 1908656/521903140
    (total drops/bytes drops) 0/0
    Class-map: Call-Control (match-any)
    615085 packets, 48926098 bytes
    30 second offered rate 84000 bps, drop rate 0 bps
    Match: ip dscp cs3 (24)
    588857 packets, 47299978 bytes
    30 second rate 81000 bps
    Match: ip dscp af31 (26)
    26228 packets, 1626120 bytes
    30 second rate 2000 bps
    Queueing
    Output Queue: Conversation 265
    Bandwidth 500 (kbps) Max Threshold 64 (packets)
    (pkts matched/bytes matched) 337953/26882724
    (depth/total drops/no-buffer drops) 0/0/0
    Class-map: Video (match-any)
    146136 packets, 82165408 bytes
    30 second offered rate 90000 bps, drop rate 0 bps
    Match: ip dscp af41 (34)
    146136 packets, 82165408 bytes
    30 second rate 90000 bps
    Queueing
    Output Queue: Conversation 266
    Bandwidth 3220 (kbps) Max Threshold 64 (packets)
    (pkts matched/bytes matched) 81687/45950190
    (depth/total drops/no-buffer drops) 0/0/0
    Class-map: class-default (match-any)
    35227089 packets, 47492000208 bytes
    30 second offered rate 87718000 bps, drop rate 14714000 bps
    Match: any
    Queueing
    Flow Based Fair Queueing
    Maximum Number of Hashed Queues 256
    (total queued/total drops/no-buffer drops) 0/5171786/0
    exponential weight: 9
    class Transmitted Random drop Tail drop Minimum Maximum Mark
    pkts/bytes pkts/bytes pkts/bytes thresh thresh prob
    0 30181523/39910255774 1297726/1944176143 3893194/5836883998 20 40 1/10
    1 0/0 0/0 0/0 22 40 1/10
    2 0/0 0/0 0/0 24 40 1/10
    3 0/0 0/0 0/0 26 40 1/10
    4 0/0 0/0 0/0 28 40 1/10
    5 0/0 0/0 0/0 30 40 1/10
    6 1213/88749 0/0 0/0 32 40 1/10
    7 0/0 0/0 0/0 34 40 1/10
    rsvp 0/0 0/0 0/0 36 40 1/10

  • Export/import character data conversion in NLS_LANG FAQ

    The [OTN NLS_LANG FAQ|http://www.oracle.com/technology/tech/globalization/htdocs/nls_lang%20faq.htm#_Toc110410566] says:
    >
    In Oracle9i the Export utility always exports user data, including Unicode data, in the character set of the database. The Import utility automatically converts the data to the character set of the target database.
    >
    But I don't think this is correct for Oracle 9i. Example:
    export NLS_LANG=AMERICAN_AMERICA.US7ASCII
    sqlplus /
    SQL*Plus: Release 9.2.0.6.0 - Production on Wed Jun 17 10:05:57 2009
    Copyright (c) 1982, 2002, Oracle Corporation.  All rights reserved.
    Connected to:
    Oracle9i Enterprise Edition Release 9.2.0.6.0 - 64bit Production
    JServer Release 9.2.0.6.0 - Production
    SQL> select * from v$version;
    BANNER
    Oracle9i Enterprise Edition Release 9.2.0.6.0 - 64bit Production
    PL/SQL Release 9.2.0.6.0 - Production
    CORE    9.2.0.6.0       Production
    TNS for HPUX: Version 9.2.0.6.0 - Production
    NLSRTL Version 9.2.0.6.0 - Production
    SQL> select * from nls_database_parameters where parameter like '%SET%';
    PARAMETER                      VALUE
    NLS_CHARACTERSET               WE8ISO8859P15
    NLS_NCHAR_CHARACTERSET         AL16UTF16
    SQL> exit
    Disconnected from Oracle9i Enterprise Edition Release 9.2.0.6.0 - 64bit Production
    JServer Release 9.2.0.6.0 - Production
    exp / owner=test file=test.dmp
    Export: Release 9.2.0.6.0 - Production on Wed Jun 17 10:06:42 2009
    Copyright (c) 1982, 2002, Oracle Corporation.  All rights reserved.
    Connected to: Oracle9i Enterprise Edition Release 9.2.0.6.0 - 64bit Production
    JServer Release 9.2.0.6.0 - Production
    Export done in US7ASCII character set and AL16UTF16 NCHAR character set
    server uses WE8ISO8859P15 character set (possible charset conversion)
    About to export specified users ...Could you please:
    - check and update NLS_LANG FAQ for Oracle 9i
    - update NLS_LANG FAQ for Oracle 10g, Oracle 11g and Data Pump.
    Thanks.
    Edited by: P. Forstmann on Jun 17, 2009 10:14 AM

    I was wrong and you are right like the OTN NLS_LANG FAQ.
    I made the following test with Oracle 9.2.0.6 and with Oracle XE:
    sqlplus test/test
    SQL*Plus: Release 10.2.0.1.0 - Production on Ven. Juin 19 10:20:23 2009
    Copyright (c) 1982, 2005, Oracle.  All rights reserved.
    Connecté à :
    Oracle Database 10g Express Edition Release 10.2.0.1.0 - Production
    SQL> create table t (x varchar2(10));
    Table créée.
    SQL> insert into t values(chr(233));
    1 ligne créée.
    SQL> select * from t;
    X
    é
    SQL> select dump(x,1017) from t;
    DUMP(X,1017)
    Typ=1 Len=1 CharacterSet=WE8MSWIN1252: e9
    SQL> exit
    Déconnecté de Oracle Database 10g Express Edition Release 10.2.0.1.0 - Production
    set NLS_LANG=AMERICAN_AMERICA.US7ASCII
    exp test/test tables=t file=t.dmp
    Export: Release 10.2.0.1.0 - Production on Fri Jun 19 10:21:54 2009
    Copyright (c) 1982, 2005, Oracle.  All rights reserved.
    Connected to: Oracle Database 10g Express Edition Release 10.2.0.1.0 - Production
    Export done in US7ASCII character set and AL16UTF16 NCHAR character set
    server uses WE8MSWIN1252 character set (possible charset conversion)
    About to export specified tables via Conventional Path ...
    . . exporting table                              T          1 rows exported
    Export terminated successfully without warnings.
    sqlplus test/test
    SQL*Plus: Release 10.2.0.1.0 - Production on Fri Jun 19 10:22:07 2009
    Copyright (c) 1982, 2005, Oracle.  All rights reserved.
    Connected to:
    Oracle Database 10g Express Edition Release 10.2.0.1.0 - Production
    SQL> drop table t;
    Table dropped.
    SQL> exit
    Disconnected from Oracle Database 10g Express Edition Release 10.2.0.1.0 - Production
    imp test/test tables=t file=t.dmp
    Import: Release 10.2.0.1.0 - Production on Fri Jun 19 10:22:23 2009
    Copyright (c) 1982, 2005, Oracle.  All rights reserved.
    Connected to: Oracle Database 10g Express Edition Release 10.2.0.1.0 - Production
    Export file created by EXPORT:V10.02.01 via conventional path
    import done in US7ASCII character set and AL16UTF16 NCHAR character set
    import server uses WE8MSWIN1252 character set (possible charset conversion)
    . importing TEST's objects into TEST
    . importing TEST's objects into TEST
    . . importing table                            "T"          1 rows imported
    Import terminated successfully without warnings.
    sqlplus test/test
    SQL*Plus: Release 10.2.0.1.0 - Production on Fri Jun 19 10:22:32 2009
    Copyright (c) 1982, 2005, Oracle.  All rights reserved.
    Connected to:
    Oracle Database 10g Express Edition Release 10.2.0.1.0 - Production
    SQL>  select dump(x,1017) from t;
    DUMP(X,1017)
    Typ=1 Len=1 CharacterSet=WE8MSWIN1252: e9

Maybe you are looking for

  • OSX 10.7.4 doesn't recognize boot disk

    My System: Model Name:          MacBook Pro   Model Identifier:          MacBookPro2,2   Processor Name:          Intel Core 2 Duo   Processor Speed:          2.33 GHz   Number of Processors:          1   Total Number of Cores:          2   L2 Cache:

  • BTE for F871

    Hello, Is there any BTE which can be triggered when a "Payment Request" is created (Tcode: F871). What is the event number?

  • Annoying popups constantly asking me to download things -help

    I've used i-antivirus and deleted infected files. And yet, everytime I use firefox, annoying popups emerge when I click on a website (say, BBC.com or Netflix.com), smaller than the main windown, usually asking me to download something, such as "flash

  • New OSX version release date - worth postponing purchase?

    I was wondering if there is any information on the date of release of new version (10.5) of OSX? Thought it could be worth postponing purchase of new imac g5, if the date is 1-2 months away. Thank you in advance.

  • Help: how to add a button to print the screen

    Hi, I am working on a set of forms. One of the feature user ask is to place a "Print Screen" button on the form so that user will be about to simply print the form screen. Any help will be greatly appreciated. Thank you in advance. Jimmy