Generic way of handling SoftKeys in Canvas (low level API)

Folks, is there a way to handle SOFTKEY presses in J2ME. I am aware of how Nokia does it, but, given a Canvas, is it possible to write a cross-device-protable code which can detect which SOFTKEY has been pressed

hello, i have had similar issues when trying to find a portable solution to using softkeys with the canvas. In the end i decided that the cross-phone support was not possible for the left and right softkeys(nokia left softkey = 6, moto left soft key =21?? or something like that?), however i did find that the fire and arrow keys are mapped the same for almost all phones and brands, instead of using the softkeys, i click once to display a pop-up menu, where the user can move up or down then select again with the fire button.. seems to work pretty well. so no softkeys?.../freddie

Similar Messages

  • Example using low level API POI for generating Excel??

    Hi
    is anyone know good URL that teaches generating Excel using low level API using Apache POI?
    The reason is because i want to set the password for either workbook/sheet in high level API but it cant be done.
    the only thing can be done using High level API is protect sheet without password.
    Any idea?
    Thanks

    There is quite a good pack of documentation that ships with HSSF. To answer your question, I have just gone to the how to page and copied the example below, hope it helps.
    short rownum;
    // create a new file
    FileOutputStream out = new FileOutputStream("workbook.xls");
    // create a new workbook
    HSSFWorkbook wb = new HSSFWorkbook();
    // create a new sheet
    HSSFSheet s = wb.createSheet();
    // declare a row object reference
    HSSFRow r = null;
    // declare a cell object reference
    HSSFCell c = null;
    // create 3 cell styles
    HSSFCellStyle cs = wb.createCellStyle();
    HSSFCellStyle cs2 = wb.createCellStyle();
    HSSFCellStyle cs3 = wb.createCellStyle();
    HSSFDataFormat df = wb.createDataFormat();
    // create 2 fonts objects
    HSSFFont f = wb.createFont();
    HSSFFont f2 = wb.createFont();
    //set font 1 to 12 point type
    f.setFontHeightInPoints((short) 12);
    //make it blue
    f.setColor( (short)0xc );
    // make it bold
    //arial is the default font
    f.setBoldweight(HSSFFont.BOLDWEIGHT_BOLD);
    //set font 2 to 10 point type
    f2.setFontHeightInPoints((short) 10);
    //make it red
    f2.setColor( (short)HSSFFont.COLOR_RED );
    //make it bold
    f2.setBoldweight(HSSFFont.BOLDWEIGHT_BOLD);
    f2.setStrikeout( true );
    //set cell stlye
    cs.setFont(f);
    //set the cell format
    cs.setDataFormat(df.getFormat("#,##0.0"));
    //set a thin border
    cs2.setBorderBottom(cs2.BORDER_THIN);
    //fill w fg fill color
    cs2.setFillPattern((short) HSSFCellStyle.SOLID_FOREGROUND);
    //set the cell format to text see HSSFDataFormat for a full list
    cs2.setDataFormat(HSSFDataFormat.getBuiltinFormat("text"));
    // set the font
    cs2.setFont(f2);
    // set the sheet name in Unicode
    wb.setSheetName(0, "\u0422\u0435\u0441\u0442\u043E\u0432\u0430\u044F " +
                       "\u0421\u0442\u0440\u0430\u043D\u0438\u0447\u043A\u0430",
                    HSSFWorkbook.ENCODING_UTF_16 );
    // in case of compressed Unicode
    // wb.setSheetName(0, "HSSF Test", HSSFWorkbook.ENCODING_COMPRESSED_UNICODE );
    // create a sheet with 30 rows (0-29)
    for (rownum = (short) 0; rownum < 30; rownum++)
        // create a row
        r = s.createRow(rownum);
        // on every other row
        if ((rownum % 2) == 0)
            // make the row height bigger  (in twips - 1/20 of a point)
            r.setHeight((short) 0x249);
        //r.setRowNum(( short ) rownum);
        // create 10 cells (0-9) (the += 2 becomes apparent later
        for (short cellnum = (short) 0; cellnum < 10; cellnum += 2)
            // create a numeric cell
            c = r.createCell(cellnum);
            // do some goofy math to demonstrate decimals
            c.setCellValue(rownum * 10000 + cellnum
                    + (((double) rownum / 1000)
                    + ((double) cellnum / 10000)));
            String cellValue;
            // create a string cell (see why += 2 in the
            c = r.createCell((short) (cellnum + 1));
            // on every other row
            if ((rownum % 2) == 0)
                // set this cell to the first cell style we defined
                c.setCellStyle(cs);
                // set the cell's string value to "Test"
                c.setEncoding( HSSFCell.ENCODING_COMPRESSED_UNICODE );
                c.setCellValue( "Test" );
            else
                c.setCellStyle(cs2);
                // set the cell's string value to "\u0422\u0435\u0441\u0442"
                c.setEncoding( HSSFCell.ENCODING_UTF_16 );
                c.setCellValue( "\u0422\u0435\u0441\u0442" );
            // make this column a bit wider
            s.setColumnWidth((short) (cellnum + 1), (short) ((50 * 8) / ((double) 1 / 20)));
    //draw a thick black border on the row at the bottom using BLANKS
    // advance 2 rows
    rownum++;
    rownum++;
    r = s.createRow(rownum);
    // define the third style to be the default
    // except with a thick black border at the bottom
    cs3.setBorderBottom(cs3.BORDER_THICK);
    //create 50 cells
    for (short cellnum = (short) 0; cellnum < 50; cellnum++)
        //create a blank type cell (no value)
        c = r.createCell(cellnum);
        // set it to the thick black border style
        c.setCellStyle(cs3);
    //end draw thick black border
    // demonstrate adding/naming and deleting a sheet
    // create a sheet, set its title then delete it
    s = wb.createSheet();
    wb.setSheetName(1, "DeletedSheet");
    wb.removeSheetAt(1);
    //end deleted sheet
    // write the workbook to the output stream
    // close our file (don't blow out our file handles
    wb.write(out);
    out.close();

  • Can I use textfield in an application with low level api?

    Hello !
    I am developing a mobile application using the low level api.
    I have to use it, because I do a lot of coloring.
    But I want to use high level textfields as well. Is there any way
    to include them in my application without using forms?
    Does any one know, why J2ME designer's have limited the access
    to the TextField's paint()-method? That's why I cannot inherit
    from TextField and overwrite the paint() as I would like to do.
    Thanks.

    Unfortunately no. You cannot mix both high and low level on the same screen. You'll probably have to draw it out yourself and capture user input, unless you're willing to use textboxes.
    Or.. i suppose you can try using customItems for all your "coloring" instead?

  • HIgh Level API's Vs Low Level API's

    Hi,
    Is it advisable to use High Level API's or Low Level API's for developing a Mobile Application?
    What will be the advantages/disadvantages of using High Level API's over Low Level API's or vice versa?
    Kindly advise.

    Is it advisable to use High Level API's or Low Level API's for developing a Mobile Application?What have you used so far?
    What will be the advantages/disadvantages of using High Level API's over Low Level API's or vice versa?What are your own thoughts on this?
    {color:#0000ff}http://catb.org/~esr/faqs/smart-questions.html{color}
    db

  • How to handle softkey events

    hi all,
    The softkeys varies form device to device.so kidly say some genric way to handle softkey events.
    Thanks in advance
    Badri

    The only "generic" way to handle softkeys is to use Commands.
    shmoove

  • Low level network info

    I'm coding a cellular phone program. It needs low level network info, idealy it would be a CellID that it is connected to.
    The main problem is to know if two phones ar in the same area without using location ('cos it is a relatively new api and here in argentina time will past until location phones are here ).
    The simplest thing i imagine is to know the cellid where the phone is connected and match in a server what phones are in the same cellid. I was looking for a low level api or something to use AT commands in the phone but i can't find nothing.
    If you know any way to acces that info or the AT commands please respond.
    If you know other way to know if two phones are in the same area, please let my know.
    When i say "same area", i mean like 20-30 milles (50km) from each other, more or less, it dont need to be accurate.
    Thanks in advance
    Losky

    This seems to be a bug. When a transformation (to 3785, 3857) is involved it seems to query the whole table to build up the cache.  Will investigate further to see if caching=none makes a difference.

  • Handle model nodes in a generic way

    Hello,
    Is there any way to manage model nodes and model attributes in somewhat generic way?
    I am trying to build a generic component which can handle generic model objects execution. For example, using a generic method executeModelObject(IWDNode node) who initilialize a model node, populates it and execute it, but without using specifical custom classes generated by WebDynpro Framework. I think, that the code must be specifical to a particualr model node.
    Any help?

    David,
    You can switch off 'TypedAccessRequired' property for any node at design time and then these nodes can be accessed with generic API - you need to know the name of node and element index though.
    See [this presentation|http://www.google.com/url?sa=t&source=web&ct=res&cd=7&url=http%3A%2F%2Fwww.bus.ucf.edu%2Frhightower%2Fism4154%2FNotes%2FWD%2FContext%2520Programming.ppt&ei=L1tNSa-BPJiU8wT47uiZDw&usg=AFQjCNGlcTAwGMrxidkcothQOl-U6Kf6rw&sig2=3M03WTUxex4EBWJCeCfIIA] (Slides 6 and 7 in particular).
    Hope this helps.
    Vishwas.

  • Best way to handle optional elements in XSLT

    Hi All,
    I'v wondering if there are any better ways to handle optional fields in XSLT.
    For example if my Person Type has 5 fields and 3 of them are optional:
    <xs:complexType name="PersonType">
    <xs:element name="PersonID" min="1"/>
    <xs:element name="PIN" min="1"/>
    <xs:element name="FirstName" min="0"/>
    <xs:element name="MiddleName" min="0"/>
    <xs:element name="LastName" min="0"/>
    </xs:complexType>
    In my xslt file, if I do somthing like:
    <Person>
    <FirstName>
    <xsl:value-of select="PersonType/FirstName"/>
    </FirstName>
    </Person>
    I might get an empty FirstName element in my output xml file. (like <FirstName/>) This has caused some very weired behavour with Oracle v2 parser because optional means the node may not be present in the xml not the node is there but empty.
    The only way I can think of solving this is to put <xsl:if test> around every single optional element to test if it exist in the source file. If it does then generate the element in the destination file. But in this way, if the source file is very complicated, then you would have to wrap each element with a <xsl:test> which is very ugly to me.
    Do anyone know if there is a more generic way that we can solve this issue?
    Thanks,
    Larry

    Dharmendra,
    The answer you gave is very very helpful. The only thing that I can think of here is that it might accidentally delete the fields that have been set empty deliberately.
    Maybe, this is one of the grey areas in XML Schema that I do not understand very well yet. In xml, if you define an element to be a string type and mandatory (minOccurs="1"), for exampe the following xml
    <Name></Name>
    will be valid against the schema because the Name element is present in the xml payload. However, it has no value or has a valud of empty string. I know that this is only a problem with string type. If it is boolean, then it has to be either true of false, being empty won't satisfy the schema validator. But with String it is a bit trickier, because String can be empty. Maybe this issue should be stopped when designing your xml schema by specifying the mininum length (>0) in the schema. Any suggestions would be welcomed.
    Cheers,
    Larry

  • Way to handle fact table

    hi,
    let;s say that we have a fact table with 250 millions rows of data.
    9 dimensions.
    My question is ,even we make aggregates to all possible levels,the queries in low levels are really slow(30-60 seconds).
    We use indexes in the fact table in columns where we have the joins to dimensions.
    Any ideas???
    Is it a way to have the same logical fact and make more datasources,meaning
    measures in fact
    and let;s say 2 dimensions,product and supplier.
    In one product_dim have the measures of the items begins from a to g ,
    second product_dim have the measures of the items begins from g to t ,
    and third product_dim have the measures of the items begins from u to z .
    So if the queries wants data from one Dsource go from a smaller fact (with the appropriates aggregates)...
    and if the query requires data from both or all of them,Bi suite make the union all by itself...
    Any other recomondation...
    tnks

    I'd start with a trainining education session with your users discussing the merits of star schemas.
    Accept OBIEE does not handle Outerjoins that well in current guise (at least with extending the outer join syntax to where clause predicates / filters)
    Resolve any missing fact / dimension records with a '-1' dimension key in the fact and a corresponding -1 PK with 'unknown' attributes in the dimension - not everybody likes this idea but it is a tradeoff with the outer join problem.
    Change your ETL routines to create either
    a) consolidated facts where the join is already done (ie between sales orders + orders + invoices)
    b) simple indexed bridge / mapping tables that again, hold 'unknown' for missing relationships between sales orders / invoices / and negate the need for outerjoins.
    Just my 2c - a common misconception is that OBIEE 'can do everything' - Its only going to be as good / fast as your underlying database design - im sure you already know this but i've met a few power users / developers who feel they dont need to change the ETL at all - both OBIEE and ETL should compliment each other as a solution

  • (workflow question) - What is the best way to handle audio in a large Premiere project?

    Hey all,
    This might probably be suitable for any version of Premiere, but just in case, I use CS4 (Master Collection)
    I am wrestling in my brain about the best way to handle audio in my project to cut down on the time I am working on it.
    This project I just finished was a 10 minute video for a customer shot on miniDV (HVX-200) cut down from 3 hours of tape.
    I edited my whole project down to what looked good, and then I decided I needed to clean up all the Audio using Soundbooth, So I had to go in clip by clip, using the Edit in SoundBooth --> Render and Replace method on every clip. I couldn't find a way to batch edit any audio in Soundbooth.
    For every clip, I performed similar actions---
    1) both tracks of audio were recorded with 2 different microphones (2 mono tracks), so I needed only audio from 1 track - I used SB to cut and paste the good track over the other track.
    2) amplified the audio
    3) cleaned up the background noise with the noise filter
    I am sure there has to be a better workflow option than what I just did (going clip by clip), Can someone give me some advice on how best to handle audio in a situation like this?
    Should I have just rendered out new audio for the whole tape I was using, and then edit from that?
    Should I have rendered out the audio after I edited the clips into one long track and performed the actions I needed on it? or something entirely different? It was a very slow, tedious process.
    Thanks,
    Aza

    Hi, Aza.
    Given that my background is audio and I'm just coming into the brave new world of visual bits and bytes, I would second Hunt's recommendation regarding exporting the entire video's audio as one wav file, working on it, and then reimporting. I do this as one of the last stages, when I know I have the editing done, with an ear towards consistency from beginning to end.
    One of the benefits of this approach is that you can manage all audio in the same context. For example, if you want to normalize, compress or limit your audio, doing it a clip at a time will make it difficult for you to match levels consistently or find a compression setting that works smoothly across the board. It's likely that there will instead be subtle or obvious differences between each clip you worked on.
    When all your audio is in one file you can, for instance, look at the entire wave form, see that limiting to -6 db would trim off most of the unnecessary peaks, triim it down, and then normalize it all. You may still have to do some tweaking here and there, but it gets you much farther down the road, much more easily.Same goes for reverb, EQ or other effects where you want the same feel throughout the entire video.
    Hope this helps,
    Chris

  • What's the best way to handle this?

    I'm not sure what APIs/setup to use for this situation:
    A company wants to store data projects they do for clients. Each year, the data fields are set (as a result of gov't requirements) and they won't change for any client project for that year. however, the fields required can (and usually do) change every year. So things they require this year, might not be needed the next year and new fields might be introduced.
    While there are likely to be many common fields from year to year, there's no way to guarantee which ones will remain consistent. They also want to be able to do searches on the data and fields, for projects within a year and across years.
    What's the best framework/API/configuration to handle this? EJB? Simple JDBC? If so, how should the database be handled? Won't it have to constantly create new fields in a table? Or is there another way to handle this?
    What's the best way from a "clean architecture" standpoint?

    dang, I really have to start over? I finally got all this stuff working again.  well, hopefully it won't be as big a pain this time since the data won't be coming from a different machine.   After completing the Migration Assistant process, I had to reinput a bunch of serial numbers for apps, reinstall print and mouse drivers, etc...  I've finally got the new machine up and running smoothly and now I gotta start over? Sigh.
    I was hoping that either I could rename the current account after deleting the other one, or just move everything from one account to the other and then delete the 'RJM' account.
    ok, so it sounds like here are the steps I need to take:
    - make another full cloned backup of this current machine in Super Duper
    - reboot this machine using the advice in the first post, wipe everything clean and reinstall the OS
    - create a new account like 'user1' and re-do software update (which is like 2.5 gig worth of stuff) and takes like an hour even on a high speed connection
    - then re-do the migration assistant process to the properly named account
    - then delete the 'user1' account
    does that sound right?

  • What are the different ways to handle deadlocks?

    Hi,
    May I know what are the ways to solve a deadlock problem?
    Currently, I have the following code to catch the exception:
    catch (XmlException ex)
                   try
                        ex.printStackTrace();
                        txn.abort();
                   } catch (DatabaseException DbEx)
                        System.err.println("txn abort failed.");
                   }and the resulting error is:
    com.sleepycat.dbxml.XmlException: Error: DB_LOCK_DEADLOCK: Locker killed to resolve a deadlock, errcode = DATABASE_ERROR
    Any other more efficient way to handle deadlock?
    Or better ways to prevent deadlock from happening?
    I am using this environment config
    EnvironmentConfig envConf = new EnvironmentConfig();
                   envConf.setAllowCreate(true); // If the environment does not exits,
                   // create it.
                   envConf.setInitializeCache(true); // Turn on the shared memory
                   // region.
                   // envConf.setCacheSize(25 * 1024 * 1024); // 25MB cache
                   envConf.setInitializeLocking(true); // Turn on the locking
                   // subsystem.
                   envConf.setInitializeLogging(true); // Turn on the logging
                   // subsystem.
                   envConf.setTransactional(true); // Turn on the transactional
                   // subsystem.
                   // envConf.setRunRecovery(true); //Turn on run recovery
                   // envConf.setTxnNoSync(true); // Cause BDB XML to not synchronously
                   // force any log data to disk upon transaction commit
                   envConf.setLogInMemory(true); // specify in-memory logging
                   envConf.setLogBufferSize(60 * 1024 * 1024); // set logging size.
                   // envConf.setTxnWriteNoSync(true); //method. This causes logging
                   // data to be synchronously written to the OS's file system buffers
                   // upon transaction commit.
                   // envConf.setThreaded(true); //default by Java that threaded = true
                   // envConf.setMultiversion(true);
                   envConf.setLockDetectMode(LockDetectMode.DEFAULT); // Reject a
                   // random lock
                   // requestThanks in advance for any help!
    :)

    Hi Vyacheslav,
    here is the code:
    package ag;
    import com.sleepycat.db.DatabaseException;
    import com.sleepycat.db.Environment;
    import com.sleepycat.db.EnvironmentConfig;
    import com.sleepycat.db.LockDetectMode;
    import com.sleepycat.dbxml.XmlContainerConfig;
    import com.sleepycat.dbxml.XmlDocumentConfig;
    import com.sleepycat.dbxml.XmlException;
    import com.sleepycat.dbxml.XmlManager;
    import com.sleepycat.dbxml.XmlContainer;
    import com.sleepycat.dbxml.XmlDocument;
    import com.sleepycat.dbxml.XmlManagerConfig;
    import com.sleepycat.dbxml.XmlTransaction;
    import com.sleepycat.dbxml.XmlUpdateContext;
    import inter.DBInterface;
    import java.io.*;
    import java.util.Properties;
    import cp.CheckPointer;
    public class SaveMessageinDB implements DBInterface
         Environment myEnv;
         XmlManager myManager;
         XmlContainer myContainer;
         XmlTransaction txn;
         XmlContainerConfig cconfig;
         Properties properties;
         // CheckPointer cp;
         int Counter;
         public SaveMessageinDB()
              try
                   properties = new Properties();
                   properties.load(ClassLoader
                             .getSystemResourceAsStream("Aggregator.properties"));
                   setXmlEnvrionment();
                   setXmlManager();
                   setXmlContainer();
                   // cp = new CheckPointer(myEnv);
                   // cp.start();
                   // System.out.println("Checkpointer started....");
                   Counter = 0;
              } catch (Exception ex)
                   ex.printStackTrace();
         public void saveMessage(String docName, String content) throws Exception
              addXMLDocument(docName, content);
         public void setXmlEnvrionment()
              try
                   File envHome = new File(properties.getProperty("DATABASE_LOCATION"));
                   EnvironmentConfig envConf = new EnvironmentConfig();
                   envConf.setAllowCreate(true); // If the environment does not exits,
                   // create it.
                   envConf.setInitializeCache(true); // Turn on the shared memory
                   // region.
                   envConf.setCacheSize(100 * 1024 * 1024); // 100MB cache
                   envConf.setInitializeLocking(true); // Turn on the locking
                   // subsystem.
                   envConf.setInitializeLogging(true); // Turn on the logging
                   // subsystem.
                   envConf.setTransactional(true); // Turn on the transactional
                   // subsystem.
                   // envConf.setRunRecovery(true); // Turn on run recovery
                   // envConf.setTxnNoSync(true); // Cause BDB XML to not synchronously
                   // force any log data to disk upon transaction commit
                   envConf.setLogInMemory(true); // specify in-memory logging
                   envConf.setLogBufferSize(60 * 1024 * 1024); // set logging size.
                   // envConf.setTxnWriteNoSync(true);
                   // This causes logging
                   // data to be synchronously written to the OS's file system buffers
                   // upon transaction commit.
                   envConf.setMultiversion(true); //Turn on snapshot isolation
                   envConf.setLockDetectMode(LockDetectMode.DEFAULT); // Reject a
                   // random lock
                   // request
                   // myEnv = new Environment(envHome, null); //To adopt Environment
                   // already set by others
                   myEnv = new Environment(envHome, envConf);
                   System.out.println("Environment created...");
              } catch (Exception ex)
                   ex.printStackTrace();
         // All BDB XML programs require an XmlManager instance.
         // Create it from the DB Environment, but do not adopt the
         // Environment
         public void setXmlManager()
              try
                   XmlManagerConfig mconfig = new XmlManagerConfig();
                   mconfig.setAllowAutoOpen(true);
                   mconfig.setAdoptEnvironment(true);
                   mconfig.setAllowExternalAccess(true);
                   myManager = new XmlManager(myEnv, mconfig);
                   // myManager = new XmlManager (mconfig);
                   System.out.println("Manager created...");
              } catch (Exception ex)
                   ex.printStackTrace();
         public void setXmlContainer()
              try
                   cconfig = new XmlContainerConfig();
                   cconfig.setNodeContainer(true);
                   cconfig.setIndexNodes(true);
                   cconfig.setTransactional(true); // set transaction need an
                   // cconfig.setAllowValidation(false);
                   // environment
                   // cconfig.setReadUncommitted(true); // This container allow
                   // uncommitted read (able to read dirty data and not set a deadlock
                   // cconfig.setMultiversion(true);
                   myContainer = myManager.openContainer(properties
                             .getProperty("DATABASE_LOCATION")
                             + properties.getProperty("CONTAINER_NAME"), cconfig);
                   System.out.println("Container Opened...");
              } catch (XmlException XmlE)
                   try
                        myContainer = myManager.createContainer(properties
                                  .getProperty("DATABASE_LOCATION")
                                  + properties.getProperty("CONTAINER_NAME"), cconfig);
                        System.out.println("Container Created...");
                   } catch (Exception e)
                        e.printStackTrace();
              } catch (Exception ex)
                   ex.printStackTrace();
         public void addXMLDocument(String docName, String content)
              try
                   txn = myManager.createTransaction(); // no need to create
                   // transaction. auto commit
                   // by the environment
                   XmlDocumentConfig docConfig = new XmlDocumentConfig();
                   docConfig.setGenerateName(true);
                   docConfig.setWellFormedOnly(true);
                   myContainer.putDocument(txn, docName, content, docConfig);
                   // commit the Transaction
                   txn.commit();
                   System.out.println("documents added.....");
                   Counter++;
                   System.out.println("Document no: " + Counter);
                   txn.delete();
              } catch (XmlException ex)
                   try
                        System.out.println("Occuring in addXMLDocument");
                        ex.printStackTrace();
                        txn.abort();
                   } catch (DatabaseException DbEx)
                        System.err.println("txn abort failed.");
         public void cleanup()
              try
                   if (myContainer != null)
                        myContainer.close();
                   if (myManager != null)
                        myManager.close();
                   if (myEnv != null)
                        System.out.println("All cleaned up done..in sm");
                        myEnv.close();
              } catch (Exception e)
                   // ignore exceptions in cleanup
    }Thanks!

  • Ways to handle large volume data (file size = 60MB) in PI 7.0 file to file

    Hi,
    In a file to file scenario (flat file to xml file), the flat file is getting picked up by FCC and then send to XI. In xi its performing message mapping and then XSL transformation in a sequence.
    The scenario is working fine for small files (size upto 5MB) but when the input flat file size is more then 60 MB, then XI is showing lots of problem like (1) JCo call error or (2) some times even XI is stoped and we have to strat it manually again to function properly.
    Please suggest some way to handle large volume (file size upto 60MB) in PI 7.0 file to file scenario.
    Best Regards,
    Madan Agrawal.

    Hi Madan,
    If every record of your source file was processed in a target system, maybe you could split your source file into several messages by setting up this in Recordset Per Messages parameter.
    However, you just want to convert you .txt file into a .xml file. So, try firstly to setting up
    EO_MSG_SIZE_LIMIT parameter in SXMB_ADM.
    However this could solve the problem in Inegration Engine, but the problem will persit in Adapter Engine, I mean,  JCo call error ...
    Take into account that file is first proccessed in Adapter Engine, File Content Conversion and so on...
    and then it is sent to the pipeline in Integration Engine.
    Carlos

  • Best way to handle text files in OD10g

    We have a requirement to store reports in text format into a database field, to be able to view the reports, and to print them if desired using Forms 10g. What is the best way to handle this?
    - define the field in the database as clob or blob?
    - if CLOB is the choice, what tools to use to upload CLOBs to the database (since webutil transfer is for blob only)?
    - in Forms 10g, can one use the Forms data type LONG for CLOB?
    - can you do Forms search on clob and blob fields?
    - how can reports that are stored in fields be viewed without first downloading to the client workstation?
    - in Forms 10g, what is the best way to view text files residing in local PCs: "host notepad myFile"?
    Thanks much for your reply!
    gk

    Take a deep breath. Relax. All is fine.
    iDVD does not look at the size of your video file, it looks at the length. iDVD can accomodate up to 2 hours of movie
    iDVD gives you different options depending on the length of your movie. Although I won't agree with your friend about reducing the length of your movie to 15 minutes, if you could trim out a few minutes to get it under an hour that setting in iDVD (Best Performance though the new version may have renamed it) gives you the best quality. Still, any iDVD setting will give you good quality even at 64 minutes
    In FCE export as Quicktime Movie NOT any flavour of Quicktime Conversion. Select chapter markers if you have them. If everything is on one system unchecked the Make Movie Self Contained button. Drop the QT file into iDVD

  • Best way to handle multiple currencies

    I have a requirement that users should be able to report against an OLAP cube in a currency of their choice (from a list of about 20) and was wondering what the best way to handle this might be.
    One option would be to have a currency dimension containing the list of valid currencies and then to pre-calculate measures in each of the currencies and store them in the cube. However the downside of this is that the resultant cube would be 20 times larger than a cube in a single currency, take longer to maintain etc. I could of course partition the cube by currency to improve reporting performance since users would only report in one currency at a time.
    Another alternative would be to dynamically calculate the measures based on exchange rates - I guess this could either be done in the cube itself or as part of the reporting code. However since exchange rates are daily, this would obvioulsy prevent me from aggregating data up the time dimension (all measures are at the day level).
    Is there any standard way of doing this and what are the pro's and con's?
    Thanks,
    Chris

    Sorry - messed up - I should have posted this in the OLAP forum.....

Maybe you are looking for

  • Sky mobile tv

    Just got the n97 mini and cant connect to sky mobile.My internet works and bbc iplayer works but the app for sky mobile fails to connect. I got told by a lad in phones 4 u that he had the same problem on his x6,that some times it works and other time

  • Table name: Infopackage Processing information

    Hi gurus! Does anybody know which table I have the information of the Infopackage tab "processing"? I need to know all infopackages that are sending data only to PSA. Thanks in advance, Silvio Messias.

  • I need to add a device to my account

    Please help how do I add another device to my Find My Iphone App? 

  • Budget Report for Project does not work

    Can somone help make sense of this inbuilt report ; Fnancial Reports/Budget Reports/Budget Report - on the Selection Criteria I put under Project - MO project. The result is the Report is that it Shows me the ACTUAL sales for MO project but the BUDGE

  • Gargeband file into kenote slide by slide

    I created a voice track for one slide in a kenote presentation. As instructed, I opened the media window, found the file, and dragged the file onto the slide pallatte. The file, however, continues to bounce back and will not import. Using the inspect