RAS Java SKD, parameters size limit with ParameterFieldController?

Hello,
I use the RAS SDK to generate a report and save it to PDF and XML files from a java web application (I'm new to Crystal and its SDK but it works great!). I'm using parameters that I set using ParameterFieldController.setCurrentValue()... everything's allright. Now, I need to add a new parameter (a list of Strings) and the user wants my application to be able to handle thousands of elements in that list... so I'm wondering... what's the limit? is there a limit?
I've not found this information into the SDK documentation or the ParameterFieldController's javadoc...
Can anyone help me on this one?
Thanks!

There's no pre-defined limit. 
But note that if you're pushing all that list of data to the RDBMS layer, the Crystal will push all that list of data to the RDBMs layer.  No guarantees of what's going to happen here.
Sincerely,
Ted Ueda

Similar Messages

  • Hitting the 32k size limit with Keyword Expansion in packages

    Hi!
    I am hitting the 32k size limit with Keyword Expansion (KE). It is hardcoded in the procedure jr_keyword.analyze_mlt.
    Are there any plans to get rid of this limit, so package bodies with size > 32000 bytes can be expanded?

    Well, I am making progress. With a combination if utl_tcp.get_line() - to trap the header and utl_tcp.get_text - to get the data at 32K chunks - which I then put into a CLOB or VARRAY of VARCHAR2; I can get as much data as is sent.
    The problem now is speed.
    It takes over 60 seconds to get 160K (5 32K chunks) of data when I put it into the VARRAY of VARCHAR2, and it takes even longer if I use dbms_lob.write() and dbms_lob.writeappend() to store the data.
    Am I doing something wrong? Is there another way?
    Thank You for any Help.
    Shannon

  • Rwcgi60 parameters size limit

    Hi all,
    Im wondering if theres a size limit to parameters passed from a web page to reports using rwcgi60.
    With the tests I did using NT parameters with more than 981 characters causes the following error in the CGI execution:
    Oracle Reports Server CGI - Unable to communicate with the Reports Server.
    Im using reports6i version 6.0.8.11.3
    I tried to use it in a Linux box with the last patch but I got the same error.
    Is this a non documented bug?

    hello,
    you might run into the restriction of 255 characters when passing the parameters on the URL. do you see the parameters in your URL after you hit submit in your form ? if yes, you are using GET to pass the parameters. use POST instead that should solve the problem.
    regards,
    the oracle reports team

  • FILE SIZE LIMIT WITH UTL_FILE?

    Hello all,
    is there any file size limit using utl_file on SUSE LINUX 5.3
    (RDBMS 8.0.5)?
    We have created a file with UTL_FILE.FOPEN, put some line into it
    (UTL_FILE.PUT_LINE), and close the file (UTL_FILE.FCLOSE).
    The file size is never greater than 32 kB.
    On 7.3.4 the same procedure works fine with files larger than 1
    MB.
    Thx
    R. Petuker
    null

    Robert Petuker (guest) wrote:
    : Hello all,
    : is there any file size limit using utl_file on SUSE LINUX 5.3
    : (RDBMS 8.0.5)?
    : We have created a file with UTL_FILE.FOPEN, put some line into
    it
    : (UTL_FILE.PUT_LINE), and close the file (UTL_FILE.FCLOSE).
    : The file size is never greater than 32 kB.
    : On 7.3.4 the same procedure works fine with files larger than 1
    : MB.
    : Thx
    : R. Petuker
    Robert,
    There is no limit on the size of the file created using the
    utl_file package. It could be the limitation of the shell
    above which you are working on. You may verify the filesize
    limit of the shell and try it again.
    Maturi
    null

  • Is there a size limit with the upload method ?

    Hi,
    I wrote an air application (with js and html) to upload files from local disc to a webserver, using a php script to uplad the file. When testing, every thing is ok wonly with files less than 8 mb, if I try to upload a 8mb file or more, there is no error message, but the file is not uploaded !!
    Is there any file size restriction in air ?
    I configured my php.ini :
    upload_max_filesize = 100M
    well but the strange thing is that phpmyadmin tells me that the maxsize for imported files is 8mb...

    In fact there is no limit, the problem is with my php server, the size was limited by the php.ini post_max_size parameter !!!
    Have a look on this :
    http://www.flashcommander.org/fc/download

  • How do I change the attachment size limit in Calendar Server 6.3, UWC, IWC?

    How do I properly increase or decrease the attachment size limit with Calendar Server and all supported user interfaces to it such as WCAP, UWC (Communications Express) and IWC (Convergence)? From my experience with the Outlook Connector, there seems to be some limit imposed by cshttpd on the size of a file upload (I believe I actually got an HTTP error code back on the wcap request indicating something was too big, sorry I don't have it handy, I'll have to re-test). Additionally, it seems UWC imposes additional limits (Example: http://docs.sun.com/app/docs/doc/819-4440/6n6jfgcjh?l=en&a=view&q=fileSizeHardLimit) but I can't seem to get those to work at all. I found many different web.xml related to UWC and I'm not sure which one to change, I tried a couple but had no success because UWC would always report this error if I uploaded between 4-5 megs: com.iplanet.jato.util.WrapperRuntimeException
    Root cause = [java.io.IOException: Request cancelled because file input field
    "importFile" size is over the configurable limit of 4194304 bytes; see filter init
    parameter fileSizeHardLimit]
    And it would complain about requestSizeLimit I think if it was over 5 megs, claiming that limit was 5242880. IWC gives a generic error when the upload is too big and rejects it.
    I fear that a 4 meg limit will be too imposing and of limited value, so I would either like to raise it, or consider lowering it to 0 bytes so attachments cannot be used at all. I have been looking high and low for information on how to do this and all I can find is the UWC examples. I plan to support the Outlook Connector, UWC, and IWC so the limits should ideally be the same across each. Some of the Exchange data we wish to import does have some attachments so it would be good to continue support for that. I did see some other posts about quota RFEs but at this point I am not concerned about the disk consumption. Can anyone help? Thanks. Please let me know if there is any more information I can provide. I am running SCS6u1 on Solaris 10 SPARC.

    Fred@egr wrote:
    Thanks!!! This is working with IWC and I am pretty sure it will work with Outlook. I didn't think to look at config options for mshttpd since I don't have it installed and ics.conf doesn't list the http.service.maxmessagesize and service.http.maxpostsize by default.http.service.maxmessagesize is only relevant to mshttpd, not cshttpd. service.http.maxpostsize applies to both.
    UWC is still limiting me though; I'm sure I can reconfigure UWC if I just know which file to edit and if I need to redeploy anything. I'm using the same install paths as the SCS6 Single Host example and I'm not sure which the "uwc-deployed-path" is supposed to be. Again, thanks.If you have deployed UWC/CE to Application Server you would edit the following file and restart application-server:
    /opt/SUNWappserver/domains/domain1/generated/xml/j2ee-modules/Communications_Express/web.xml
    e.g.
      <filter>
        <filter-name>MultipartFormServletFilter</filter-name>
        <filter-class>com.sun.uwc.calclient.MultipartFormServletFilter</filter-class>
        <init-param>
          <param-name>fileSizeHardLimit</param-name>
          <param-value>15485760</param-value>
        </init-param>
        <init-param>
          <param-name>requestSizeLimit</param-name>
          <param-value>15485760</param-value>
        </init-param>
        <init-param>
          <param-name>fileSizeLimit</param-name>
          <param-value>15485760</param-value>
        </init-param>
      </filter>Regards,
    Shane.

  • Value Mapping - size limit

    Hi all,
    We got a doubt regarding value mapping,
    Is there any size limit for value mapping tables that we are creating in Configuration?
    Thanks and regards,
    sasi

    Hi,
    since the valueMapping will be executed on Java Stack, the size limit will rely on your JVM memory setting (Heap size, basically).
    But the memory request will be from your payload. As a rule of thumb, the memory requirement will be 5-10 times of your payload size.
    Cheers,
    Aaron

  • Size limit to jstrings passed as parameters?

    I have written a JNI bridge to enable faster Xalan XSLT processing. The pure Xalan-J code is just too slow on large files, and Xalan-C++ handles the same files much quicker, so I wrote a bridge which sends an inputXML and xslt string to Xalan-C++, and receives a string of the transformedXML back.
    My code works great until it processes a big file, and then the entire inputXML string never seems to get into the nativemethod. I can print it out in its entirety on the java side just prior to calling the native method, but printing in out immediately inside the native method stops partway through the string and then my java app just keeps using almost all of my processor indefinitely. I am wondering if this is some kind of jstring size limit I am hitting, or if I have a memory allocation problem that is just slowing the machine down a bunch. Here is the relevant code:
    Java:
    private native String getTransform(String input, String stylesheet);
    public void transform(Source xmlSource, Result outputTarget) throws TransformerException {
    System.err.println("Starting Transform at "+ System.currentTimeMillis());
    System.err.println("Doc:\n"+NativeSource.getSourceAsString(xmlSource));
    String output = getTransform(NativeSource.getSourceAsString(xmlSource), stylesheet);
    System.err.println("Finished Transform at "+ System.currentTimeMillis());
    C++:
    JNIEXPORT jstring JNICALL Java_com_endeca_edf_adapter_plugins_NativeTransformer_getTransform
    (JNIEnv *pEnv, jobject obj, jstring input, jstring xslt) {
    jclass cls = pEnv->GetObjectClass(obj);
    const char *theInputDocument = pEnv->GetStringUTFChars(input, 0);
    printf("%s", theInputDocument);
    any ideas? Thanks!

    I have changed my code to split large strings into arrays and then recreate the entire string inside the C++. This still works with small files, but large ones (with base-64 encoding) don't work, just churn away forever. I made sure to set the break size real small (1096 chars) so even the small file gets broken up. Here is the code:
    Java:
    * Native method using xalan to transform xml
    private native String getTransform(String[] input, String stylesheet);
    public void transform(Source xmlSource, Result outputTarget) throws TransformerException {
    log.info("Splitting Input at "+ System.currentTimeMillis());
    int size = 0;
    String input = NativeSource.getSourceAsString(xmlSource);
    //log.info("Input XML:\n"+input);
    int numParts = (int)Math.ceil(input.length()/MAX_JNI_STRING_LENGTH);
    log.info("Number of splits= "+numParts);
    String[] inputArray = new String[numParts];
    for(int i=0; i<numParts-1; i++) {
    inputArray[i] = input.substring(size, size+MAX_JNI_STRING_LENGTH);
    size += MAX_JNI_STRING_LENGTH;
    inputArray[numParts-1] = input.substring(size);
    log.info("Starting Transform at "+ System.currentTimeMillis());
    String output = getTransform(inputArray, stylesheet);
    log.info("Finished Transform at "+ System.currentTimeMillis());
    //log.info("Transform is\n"+output);C++:
    JNIEXPORT jstring JNICALL Java_com_endeca_edf_adapter_plugins_NativeTransformer_getTransform
    (JNIEnv *pEnv, jobject obj, jobjectArray inputArray, jstring xslt) {
    //jclass cls = pEnv->GetObjectClass(obj);
    jboolean copy = JNI_FALSE;
    printf("%s", "\ninside\n");
    std::stringstream inputStr;
    for(int i=0; i<pEnv->GetArrayLength(inputArray); i++) {
         jstring inputPart = (jstring)pEnv->GetObjectArrayElement(inputArray, i);
         inputStr << pEnv->GetStringUTFChars(inputPart, &copy);
    char * theInputDocument = new char[strlen(inputStr.str().c_str())];
    strcpy(theInputDocument, inputStr.str().c_str());
    // Below wasn't working on big files, so I had to break up the string
    //const char *theInputDocument = pEnv->GetStringUTFChars(input, &copy);
    if (theInputDocument == NULL) {
         return NULL; // Could be out of memory
    printf("%s", inputStr.str().c_str());
    printf("%s", theInputDocument);Seems like I am either misunderstanding the size issue, or a jobjectArray also has size limits, or something else is going on. Help?

  • Size limit for flar file with Sol 10 WANBOOT

    There seems to be some kind of a size limit for flash archive files used by wanboot. When I try to wanboot jumpstart, I get the following:
    Processing profile
    - Opening Flash archive
    ERROR: HTTP server returned an invalid archive file size: <-1256985017> bytes
    ERROR: Invalid HTTP headers were returned from the server
    ERROR: Flash installation failed
    Solaris installation program exited.
    The flash archive is about 3GB. I'm doing the WANBOOT using the "boot cdrom" work-around, because the V100 I'm jumpstarting doesn't support the newer OBP parameters.
    If create a new smaller flash archive like so:
    mv sol10.flar sol10.flar.old
    dd if=sol10.flar.old of=sol10.flar bs=1024K count=1024
    cutting the flar down to 1GB. Then I do the wanboot. I don't get the error, and the jumpstart proceeds. Of course it would eventually have a problem when the cpio file reached premature end.
    So is there a prescribed limit to flars with WANBOOT. I'm assuming it's 2GB, but if so, why?
    No I haven't tried creating a compressed flar: that's next.
    Thanks,
    Chip Bennett
    Laurus Technologies, Inc.
    Itasca, IL

    nope, 64-bit apache2.2 and still the same error:
    Processing profile
    - Opening Flash archive
    HTTP server returned an invalid archive file size:
    <-2116410290> bytes
    ERROR: Invalid HTTP headers were returned from the
    server
    ERROR: Flash installation failed
    Solaris installation program exited.
    apache log:
    192.168.224.112 - - [06/Jul/2006:20:52:45 +0200]
    "HEAD /flash/archives/pt_sol_10_oem.flar HTTP/1.1"
    200 -
    192.168.224.112 - - [06/Jul/2006:20:52:45 +0200] "GET
    /cgi-bin/bootlog-cgi?%3Ctime%3E+wanboot_client+ident:+
    %5BID+380565+user.panic%5D+HTTP+server+returned+an+inv
    alid+archive+file+size:+%3C-2116410290%3E+bytes
    HTTP/1.1" 200 32
    192.168.224.112 - - [06/Jul/2006:20:52:45 +0200] "GET
    /cgi-bin/bootlog-cgi?%3Ctime%3E+wanboot_client+ident:+
    %5BID+592631+user.panic%5D+Invalid+HTTP+headers+were+r
    eturned+from+the+server HTTP/1.1" 200 32
    192.168.224.112 - - [06/Jul/2006:20:52:45 +0200] "GET
    /cgi-bin/bootlog-cgi?%3Ctime%3E+wanboot_client+ident:+
    %5BID+915281+user.panic%5D+Flash+installation+failed
    HTTP/1.1" 200 32
    Any help would be much appreciated.hi. i spent a long time on this issue...and got it working reliably for more than 10 gb flars.
    there are two issues- 1) apache must be 64 bit capable, and 2) you need the latest wanboot miniroot, for solaris 9 and one for solaris 10. i worked for a different division of a major car company based in stuttgart (take a guess) than i do now, and we eventually received the support we needed from sun, through the early release program. i do not know any details about if, how, or when those miniroots are otherwise available. at some point, i can only assume they will be generally available to all, if not already.
    if someone knows more about this availability issue, i would also be interested in this issue.
    thanks, dls

  • Size limit of access database with image files stored outside database

    I have read 2GB limit on size of Access database.  We are using Access 2013 database with fields that are data type attachment and storing attached images outside the database.  Does the size of the folder that is holding these images count
    in any way toward 2GB size limit?

    Hi,
    As far as I know, Access database limit size 2GB based on itself, not related to the folder size.  Attachment data
    type is similar to attaching files to e-mail messages. We need to control the Access database size is smaller then 2GB.
    When a field is defined with the attachment data type, you can store one or more files for each record in it.
    Attachments can dramatically increase the size of the database, but since the attached file is stored as a part of the dabatbase, you are not dependent on network drives being available as you would be if you inculded a hyperlink to the file. As a matter of
    fact, feel free to backup the origainal file to other disk after you attach it to the database.
    Regards,
    George Zhao
    TechNet Community Support

  • HT5213 I am about to publish my iBook with photographs. It is interactive in that small images will go full screen when touched. Is there a size limit on my book? If so how many megs?

    I am about to publish my iBook with many photographs. It is interactive in that small images will go full screen when touched. Is there a size limit on my book? If so how many megs?

    2Gb is the max

  • In perfdatasource querying for global snapshot failed with error 'the size limit for this '

    I received  scom alerts from two win 2k8 r2 servers , hosting exchange 2010 mailbox roles , the alerts came almost in same time from both servers ,
    can I ignore those alerts
    or can someone give a me a clue how can I troubleshoot those alert , please any help would be appreciated
    In PerfDataSource, querying for Global Snapshot failed with error 'The size limit for this '
    from Ops-mgmt logs 
    Log Name:      Operations Manager
    Source:        Health Service Modules
    Date:          
    Event ID:      10104
    Task Category: None
    Level:         Error
    Keywords:      Classic
    User:          N/A
    Computer:       server 1
    Description:
    In PerfDataSource, querying for Global Snapshot failed with error 'The size limit for this ' 
    One or more workflows were affected by this.  
    Workflow name: Microsoft.Windows.Server.2008.OperatingSystem.PercentMemoryUsed.Collection 
    Instance name: Microsoft Windows Server 2008 R2 Enterprise  
    Log Name:      Operations Manager
    Source:        Health Service Modules
    Date:          
    Event ID:      10104
    Task Category: None
    Level:         Error
    Keywords:      Classic
    User:          N/A
    Computer:       server 1
    Description:
    In PerfDataSource, querying for Global Snapshot failed with error 'The size limit for this ' 
    One or more workflows were affected by this.  
    Workflow name: Microsoft.Windows.Server.2008.LogicalDisk.PercentIdle.Collection 
    Instance name:  " edb file path "
    Log Name:      Operations Manager
    Source:        Health Service Modules
    Date:          
    Event ID:      10104
    Task Category: None
    Level:         Error
    Keywords:      Classic
    User:          N/A
    Computer:       server 2 
    Description:
    In PerfDataSource, querying for Global Snapshot failed with error 'The size limit for this ' 
    One or more workflows were affected by this.  
    Workflow name: Microsoft.Windows.Server.2008.NetworkAdapter.CurrentBandwidth.Collection 
    Log Name:      Operations Manager
    Source:        Health Service Modules
    Date:          
    Event ID:      10104
    Task Category: None
    Level:         Error
    Keywords:      Classic
    User:          N/A
    Computer:   server 2   
    Description:
    In PerfDataSource, querying for Global Snapshot failed with error 'The size limit for this ' 
    One or more workflows were affected by this.  
    Workflow name: Microsoft.ForefrontProtection.FPE.Server.PerformanceCollection.RealtimeScanMessageRate

    Hi Blake , 
    Thanks for your reply , I appreciate your help  ,
    I didn't put the alert from scom console because they were same as the events ( same source )
    Health Service Modules, I didn't want to spam
    more :-)
    also the two servers encountered the issue were mailbox servers and part of same DAG , it worth mention the alert were resolved
    by Exchange 2010 Correlation Engine service 
    http://blogs.technet.com/b/kevinholman/archive/2010/10/15/clustering-the-exchange-2010-correlation-engine-service.aspx
    http://support.microsoft.com/kb/2592561
    also the Opsmgmt logs are full of waring and error event like 2023 , 21402 ,  21403 , 1207 !!
    Log Name:      Operations Manager
    Source:        HealthService
    Date:          
    Event ID:      2023
    Task Category: Health Service
    Level:         Warning
    Keywords:      Classic
    User:          N/A
    Computer:      server 1
    Description:
    The health service has removed some items from the send queue for management group "SCOM" since it exceeded the maximum allowed size of 15 megabytes.
    1- alert from console >>
    In PerfDataSource, querying for Global Snapshot failed with error 'The size limit for this '
    One or more workflows were affected by this.
    Workflow name: Microsoft.Windows.Server.2008.OperatingSystem.PercentMemoryUsed.Collection
    Instance name: Microsoft Windows Server 2008 R2 Enterprise 
    EventSourceName: Health Service Modules

  • Does Hotmail have an attachment size limit? When i attach something with more than 3mb in the mail it isn't send. It gives a resize option.

    Does Hotmail have an attachment size limit? When i attach something with more than 3mb in the mail it isn't send. It gives a resize option.

    Hi. Txs. The problem is that in my pc the limit size is 25mb. But in my ipad the mail only sends files under 3mb.

  • Input pattern or replacement parameters exceed 32K size limit

    Hi,
    Following is my code
    declare
    v_clob1 clob := empty_clob;
    v_clob2 clob := empty_clob;
    begin
    select result into v_clob2 from evex_output_copy where evex_output_uid = 1;
    v_clob1 := 'a';
    v_clob1 := replace(v_clob1, 'a', v_clob2);
    end;
    the length of v_clob2 is greater than 32K.
    It shows the following error.
    input pattern or replacement parameters exceed 32K size limit
    How can I solve this ?
    Ivan

    Try the DBMS_LOB.copy procedure.

  • S1000 Data file size limit is reached in statement

    I am new to Java and was given the task to trouble shoot a java application that was written a few years ago and no longer supported. The java application creates database files the user's directory: diwdb.properties, diwdb.data, diwdb.lproperties, diwdb.script. The purpose of the application is to open a zip file and insert the files into a table in the database.
    The values that are populated in the diwdb.properties file are as follows:
    #HSQL Database Engine
    #Wed Jan 30 08:55:05 GMT 2013
    hsqldb.script_format=0
    runtime.gc_interval=0
    sql.enforce_strict_size=false
    hsqldb.cache_size_scale=8
    readonly=false
    hsqldb.nio_data_file=true
    hsqldb.cache_scale=14
    version=1.8.0
    hsqldb.default_table_type=memory
    hsqldb.cache_file_scale=1
    hsqldb.log_size=200
    modified=yes
    hsqldb.cache_version=1.7.0
    hsqldb.original_version=1.8.0
    hsqldb.compatible_version=1.8.0
    Once the databsae file gets to 2GB it brings up the error meessage 'S1000 Data file size limit is reached in statement (Insert into <tablename>......
    From searching on the itnernet it appeared that the parameter hsqldb.cache_file_scale needed to be increased & 8 was a suggested value.
    I have the distribution files (.jar & .jnlp) that are used to run the application. And I have a source directory that was found that contains java files. But I do not see any properties files to set any parameters. I was able to load both directories into NetBeans but really don't know if the files can be rebuilt for distribution as I'm not clear on what I'm doing and NetBeans shows errors in some of the directories.
    I have also tried to add parameters to the startup url: http://uknt117.uk.infores.com/DIW/DIW.jnlp?hsqldb.large_data=true?hsqldb.cache_file_scale=8 but that does not affect the application.
    I have been struggling with this for quite some time. Would greatly appreciate any assistance to help resolve this.
    Thanks!

    Thanks! But where would I run the sql statement. When anyone launches the application it creates the database files in their user directory. How would I connect to the database after that to execute the statement?
    I see the create table statements in the files I have pulled into NetBeans in both the source folder and the distribution folder. Could I add the statement there before the table is created in the jar file in the distribution folder and then re-compile it for distribution? OR would I need to add it to the file in source directory and recompile those to create a new distribution?
    Thanks!

Maybe you are looking for