Track Data size limits?

I noticed yesterday that the Track Data (Send) part of the
LMS (Send Data) (ie the lesson data variables) must have a maximum
total limit of data allowed to be transferred to the VLE
(blackboard) and back.
I am saving text entry answers in my module each answer as a
lesson data variable. However If you fill up the first activity,
when you go back into the module and add more info or info to
another activity then nothing gets saved. I presume this is because
there is a maximum ceiling of data that can be stored in these
variables. This is very worrying as we MUST be able to store
lengthy answers for all the activities and assessment questions and
the LMS MUST store these and send them back to
the module when the user returns. otherwise we will need to
redesign the module so that we can get this working some
alternative way!
From a count, it seems there can be a maximum of 1079
characters stored before it stops you sending any more data.
Any initial thoughts or better ways we can store text entry
answers via LMS KO, SCORM and Blackboard VLE?

Custom fields won't work, unless it's his own LMS which he
can modify at
will (which it isn't, in this case). And even then, that
would totally
break the SCORM spec - so his course couldn't be used in any
other LMS.
suspend.data is the only field available for random data.
There are
others, but they are pretty limited; interaction.data can't
be read
from, objective.data has available space only in the ID
field,
student.data is 2004 only (and maybe not readable).
The issue here is definitely Authorware and the Send Data KO.
I posted
to the AwareList with the same problem months ago. The Send
Data KO tops
out at about 1080 characters for some reason.
And even if that is overcome (which it can be with custom JS
functions
in the HTML), the suspend.data field is still limited to 4096
characters.
Erik
Steve Howard **AdobeCommunityExpert** wrote:
> You should be able to create custom fields in your LMS
database that you
> can use for larger storage. I have a vague memory that
there should be
> pre-existing fields that you could use too, but it's
been years since I
> spent any time reading the specs.
>
> Steve
>
>
Erik Lord
http://www.capemedia.net
Adobe Community Expert - Authorware
http://www.adobe.com/communities/experts/
http://www.awaretips.net -
samples, tips, products, faqs, and links!
*Search the A'ware newsgroup archives*
http://groups.google.com/group/macromedia.authorware

Similar Messages

  • textarea data size limits

    We are using ColdFusion 8 and MS SQL Server 2005. We have a
    form with a <textarea> tag into which users are cutting and
    pasting large documents. When the form is saved, the text is
    truncated at 32,000 characters. We are using the POST method and
    saving to a column of ntext datatype (maximum of 1,073,741,823
    characters). We do not do any programmatic data manipulation prior
    to saving - we just update using <cfquery>. We have tried
    using various "cflsqltype" settings with the same result: we
    succesfully enter more than 32,000 characters in the
    <textarea> (confirmed with JavaScript) and then only get
    32,000 characters in the database.
    This occurs with various versions of IE, FireFox, Opera and
    Safari, so it doesn't appear to be a browser issue. We have tried
    it on Windows 2000, Windows XP Pro and Windows Vista, so it's not
    particularly OS related. We have tried it on IIS 6 and IIS 7 with
    the same results, as well as SQL Server 2000 and SQL Server 2005.
    The only thing left seems to be some data incompatibility between
    ColdFusion and SQL Server.

    The CLOB was the answer - it wasn't checked! The LongText
    Buffer setting is now at 64,000 and my test works great. We'll
    probably have to up that a little more, but we don't need
    encyclopedias stored! Thanks for your quick help.

  • HTML form size limitation

    It's about a size limit. When a HTML form is submitted, there is a size limit of how much information can be sent to the server. I'm talking about using the POST submission (as opposed to the GET submission).
    It used to be that the HTTP protocol specified a limit of 8K. The new version of the protocol now says there is no limit imposed and it is up to each server manufacturer to set its own limits.
    Can anybody help?

    So what is the data size limitation of Oracle HTML form? Does anybody know that? or if there is a document can be referenced?

  • What are the limitations in terms of data size  or performance while using csv or text file as datasource?

    <p>Also what are the limitations in terms of data size  or performance related issues while using csv or text file?</p><p>Is it the best practice to use csv , text file to use as a datasource to improve performance?</p><p>Please Advice.... </p><p>&#160;</p>

    <p>Hi,</p><p>Create Same Data Input for CSV and Text File ,Create 2 different reports one for CSV and One for Text ,run them one you have done that.</p><p>Go to Report Menu and Select Performance Information .Use the Data in that to check which one is good datasource to improve performance</p><p>Cheers</p><p>Rahul</p>

  • How to Get Around the Memo Size Limitations in CR ?

    I am Using Crystal Reports 2008, SQL Database and ASP .Net Visual Studio 2010 for Team Foundation with Crystal Viewer embedded in a web page.  All current update and patches area installed.
    Database has Memo Fields up to 164000 characters in length. Viewer show fine, but with the reports that have been designed to print this information, only seeing part of the Memo field.
    This happens with both RTF, Text and HTML formatted data from within the database field.
    I have read that there is a limitation on the size of a Memo field that Crystal Reports will print (65,534).
    I actually received an Crystal Reports error box when i try to concatenate multiple substring fields as a formula.
    Does anyone have any suggestions or ideas on a work-around ? 
    Due to legal considerations, this data has to be output as it was input, so it can't be hacked. It can be parsed and again merged  but I really donu2019t want to try and write SQL procedures to parse HTML code into readable multiple pieces based on variable length tags with large memo fields.
    Please offer any and every suggestion,
    Thanks to all  ! !
    Edited by: Ludek Uher on Oct 21, 2010 1:31 PM

    yes sir,
    already did but i didn't receive any answers. . .   Memo Field Size Limitations with Crystal Reports 2008 ?
    Thanks for your help.

  • Size limitation that can be passed to Java stored procedure

    Hello!
    I enjoy using Oracle8i these days. But I have some questions
    about Java stored procedure. I want to pass the XML data to Java
    stored procedure as IN parameter. But I got some errors when the
    data size is long. Is there any limitation in the data size that
    can be passed to Java stored procedure?
    Would you please help me ?
    This message is long, but would you please read my message?
    Contents
    1. Outline : I write what I want to do and the error message I
    got
    2. About the data size boundary: I write about the boundary
    size. I found that it depend on which calling sequence I use.
    3. The source code of the Java stored procedure
    4. The source code of the Java code that call the Java stored
    procedure
    5. The call spec
    6. Environment
    1.Outline
    I want to pass the XML data to Java stored procedure. But I got
    some errors when the data size is long. The error message I got
    is below.
    [ Error messages and stack trace ]
    java.sql.SQLException: ORA-01460: unimplemented or unreasonable
    conversion reque
    sted
    java.sql.SQLException: ORA-01460: unimplemented or unreasonable
    conversion reque
    sted
    at oracle.jdbc.ttc7.TTIoer.processError(Compiled Code)
    at oracle.jdbc.ttc7.Oall7.receive(Compiled Code)
    at oracle.jdbc.ttc7.TTC7Protocol.doOall7(Compiled Code)
    at oracle.jdbc.ttc7.TTC7Protocol.parseExecuteFetch
    (TTC7Protocol.java:721
    at oracle.jdbc.driver.OracleStatement.doExecuteOther
    (Compiled Code)
    at oracle.jdbc.driver.OracleStatement.doExecuteWithBatch
    (Compiled Code)
    at oracle.jdbc.driver.OracleStatement.doExecute(Compiled
    Code)
    at
    oracle.jdbc.driver.OracleStatement.doExecuteWithTimeout(Compiled
    Code
    at
    oracle.jdbc.driver.OraclePreparedStatement.executeUpdate
    (OraclePrepar
    edStatement.java:256)
    at oracle.jdbc.driver.OraclePreparedStatement.execute
    (OraclePreparedStat
    ement.java:273)
    at javaSp.javaSpTestMain.sample_test
    (javaSpTestMain.java:37)
    at javaSp.javaSpTestMain.main(javaSpTestMain.java:72)
    2. About the data size boundary
    I don|ft know the boundary that I got errors exactly, but I
    found that the boundary will be changed if I use |gprepareCall("
    CALL insertData(?)");|h or |gprepareCall ("begin insertData
    (?); end ;")|h.
    When I use |gprepareCall(" CALL insertData(?)".
    The data size 3931 byte ---&#61664; No Error
    The data size 4045 byte ---&#61664; Error
    Whne I use prepareCall ("begin insertData(?); end ;")
    The data size 32612 byte --&#61664;No Error
    The data size 32692 byte ---&#61664; Error
    3. The source code of the Java stored procedure
    public class javaSpBytesSample {
    public javaSpBytesSample() {
    public static int insertData( byte[] xmlDataBytes ) throws
    SQLException{
    int oraCode =0;
    String xmlData = new String(xmlDataBytes);
    try{
    Connection l_connection; //Database Connection Object
    //parse XML Data
    dits_parser dp = new dits_parser(xmlData);
    //Get data num
    int datanum = dp.getElementNum("name");
    //insesrt the data
    PreparedStatement l_stmt;
    for( int i = 0; i < datanum; i++ ){
    l_stmt = l_connection.prepareStatement("INSERT INTO test
    " +
    "(LPID, NAME, SEX) " +
    "values(?, ?, ?)");
    l_stmt.setString(1,"LIPD_null");
    l_stmt.setString(2,dp.getElemntValueByTagName("name",i));
    l_stmt.setString(3,dp.getElemntValueByTagName("sex",i));
    l_stmt.execute();
    l_stmt.close(); //Close the Statement
    l_stmt = l_connection.prepareStatement("COMMIT"); //
    Commit the changes
    l_stmt.execute();
    l_stmt.close(); //Close the Statement l_stmt.execute
    (); // Execute the Statement
    catch(SQLException e ){
    System.out.println(e.toString());
    return(e.getErrorCode());
    return(oraCode);
    4. The source code of the Java code that call the Java stored
    procedure
    public static void sample_test(int num) {
    //make data
    Patient p = new Patient();
    byte[] xmlData = p.generateXMLData(num);
    try{
    // Load the Oracle JDBC driver
    DriverManager.registerDriver(new
    oracle.jdbc.driver.OracleDriver());
    Connection m_connection = DriverManager.getConnection
    ("jdbc:oracle:thin:@max:1521:test",
    "testuser", "testuser");
    CallableStatement l_stmt =
    // m_connection.prepareCall(" CALL insertData(?)");
    m_connection.prepareCall("begin insertData(?); end
    l_stmt.setBytes(1,xmlData);
    l_stmt.execute();
    l_stmt.close();
    System.out.println("SUCCESS to insert data");
    catch(SQLException e ){
    System.out.println( e.toString());
    e.printStackTrace();
    5. The call spec
    CREATE OR REPLACE PROCEDURE insertData( xmlData IN LONG RAW)
    AS
    LANGUAGE JAVA NAME 'javaSp.javaSpBytesSample.insertData(byte[])';
    6. Environment
    OS: Windows NT 4.0 SP3
    RDBMS: Oracle 8i Enterprise Edition Release 8.1.5.0.0 for
    Windows NT
    JDBC Driver: Oracle JDBC Drivers 8.1.5.0.0.
    JVM: Java1.1.6_Borland ( The test program that call Java stored
    procedure run on this Java VM)
    null

    Iam passing an array of objects from Java to the C
    file. The total size of data that Iam sending is
    around 1GB. I have to load this data into the Shared
    memory after getting it in my C file. Iam working on
    HP-UX (64-bit). Everything works fine for around 400MB
    of data. When I try to send around 500MB of data, the
    disk utilization becomes 100%, so does my memory
    utilization and I get a "Not enough space" when I try
    to access shared memory. I have allocated nearly 2.5GB
    in my SHMMAX variable. Also, I have around 45GB of
    disk free. The JVM heap size is also at 2048MB. Where did you get the 400/500 number from? Is that the size of the file?
    What do you do with the data? Are you doing nothing but copying it byte for byte into shared memory?
    If yes then a simple test is to write a C application that does the same thing. If it has problems then it means you have an environment problem.
    If no then you are probably increasing the size of the data by creating a structure to hold it. How much overhead does that add to the size of the data?

  • Nio ByteBuffer and memory-mapped file size limitation

    I have a question/issue regarding ByteBuffer and memory-mapped file size limitations. I recently started using NIO FileChannels and ByteBuffers to store and process buffers of binary data. Until now, the maximum individual ByteBuffer/memory-mapped file size I have needed to process was around 80MB.
    However, I need to now begin processing larger buffers of binary data from a new source. Initial testing with buffer sizes above 100MB result in IOExceptions (java.lang.OutOfMemoryError: Map failed).
    I am using 32bit Windows XP; 2GB of memory (typically 1.3 to 1.5GB free); Java version 1.6.0_03; with -Xmx set to 1280m. Decreasing the Java heap max size down 768m does result in the ability to memory map larger buffers to files, but never bigger than roughly 500MB. However, the application that uses this code contains other components that require the -xMx option to be set to 1280.
    The following simple code segment executed by itself will produce the IOException for me when executed using -Xmx1280m. If I use -Xmx768m, I can increase the buffer size up to around 300MB, but never to a size that I would think I could map.
    try
    String mapFile = "C:/temp/" + UUID.randomUUID().toString() + ".tmp";
    FileChannel rwChan = new RandomAccessFile( mapFile, "rw").getChannel();
    ByteBuffer byteBuffer = rwChan.map( FileChannel.MapMode.READ_WRITE,
    0, 100000000 );
    rwChan.close();
    catch( Exception e )
    e.printStackTrace();
    I am hoping that someone can shed some light on the factors that affect the amount of data that may be memory mapped to/in a file at one time. I have investigated this for some time now and based on my understanding of how memory mapped files are supposed to work, I would think that I could map ByteBuffers to files larger than 500MB. I believe that address space plays a role, but I admittedly am no OS address space expert.
    Thanks in advance for any input.
    Regards- KJ

    See the workaround in http://bugs.sun.com/bugdatabase/view_bug.do?bug_id=4724038

  • What is the recommended way to obtain tracking data from carriers post XSI

    We currently run an old version of SAP Business Connector. We are in the process of migrating all interfaces off BC onto PI. The one remaining interface we have problems is the XSI (Express Delivery Interface) interface we have with ECC06 and UPS via the BC server. The interface works but is not stable and we would like to decommission it if we are able.
    I'm not 100% clear but it appears that XSI is no longer the recommended solution for obtaining tracking data from carriers. What is the recommend method today? We'd be happy to use a PI or ABAP solution but would prefer a standard solution that was supported by SAP and UPS.

    Using Time Machine is the simplest way to back up.
    debasencio wrote:
    They don't fit on a standard data DVD and when I try to back it up to my 500GB external hard drive it says its not formatted to files that size.
    How is the drive formatted?
    If it's DOS/FAT32, you can only size file sizes up to 4GB.
    If you are using it on a Mac only, format it Mac OS X HFS+.

  • Why is Data Size in QT inspector twice the size of the file?

    I am not asking this to save space. I am just curious. When playing a dv (captured from camera) clip, I notice that the QT inspector shows data size about twice the size of the dv clip size in file system. Is it because the dv is compressed and QT decompresses?
    Actually, on clip in QT shows 22GB which is impossible to fit on a miniDV tape.

    You'll notice that a DV Stream file has two tracks (audio and video) of the same size when viewed in QuickTime Movie Inspector or the Movie Properties window of QuickTime Player Pro.
    Obviously the audio track shouldn't be the same size as the video track. There are separated from the single stream to allow editing.

  • Input file size limition in File Reader Adapter

    Hi,
    We have a File Read Activity(using File Reader Adapter) which works on a Polling model. Whenever a input file is found in a designated directory it invokes our BPEL web service. During load testing we found out that if input file size exceeds 7 MB then we get "insufficient heap space" error. We found a reference to this size limitation in http://download.oracle.com/docs/cd/E11036_01/integrate.1013/b28994.pdf which has below lines..
    "Files that are 7MB or larger in size cannot be delivered. As an alternative, you can debatch large files (if they have multiple messages), and publish these files in messages of size less than 7 MB. This alternative is applicable only to structured files (comma-delimited or fixed position), which contain more than one message. It is not applicable to binary or XML files."
    Our customer will have the data in XML format and the size of the files will be pretty huge(running into 49, 69 MBs). Do we have any workaround here or we are left with no option but to ask the customer to split the files and load it? The later option is currently difficult one for the customer since when split there is a possibility of XML becoming invalid. Any pointers in this issue is appreciated.
    Thanks!
    Balaji

    Hi,
    BPEL can not process more than 7 MB size xml structred file. but,debatching will work for xml files also provided if the file has repeated multiple structures (means multiple repeated root elements).
    Regards,
    Dhanumjay

  • JDBC:KPRB string size limits-NEED HELP!

    Hello All.
    Please, I need your help.
    I have the problems in Oracle 9.2.0.6 with stored java and jdbc:kprb internal driver.
    I try to put string(more than 32K) into LONG-type field using following java class:
    import oracle.sql.CLOB;
    import java.sql.Connection;
    import java.sql.DriverManager;
    import java.sql.PreparedStatement;
    import java.io.Reader;
    import java.io.CharArrayReader;
    public class LongTest {
    public static void insertLong()
    throws Exception {
    Connection con = DriverManager.getConnection("jdbc:default:connection:");
    //generate large string
    StringBuffer stringBuffer = new StringBuffer();
    for (int i = 0; i < 100000; i++) {
    stringBuffer.append("qwerty");
    String st = stringBuffer.toString();
    //put large string to long-type field
    PreparedStatement pst = null;
    try {
    pst = con.prepareStatement("insert into TEST_LONG (ldata) values (?)");
    pst.setString(1, st);
    pst.execute();
    } finally {
    if (pst != null) {pst.close();}
    But I get error from jdbc:kprb about limitation of data size for this type(LONG).
    But it works well in oracle 10.2.0.1...
    Has anybody a solution of how to do this in oracle 9.2.0.6???
    thanx!

    Hi,
    THis is a known limitation in the 92 RDBMS solved in 10g.
    Kuassi http://db360.blogspot.com

  • Sys.aq$_jms_text_message type queues message size limitations

    Are there any size limitations for enqueing text messages into a sys.aq$_jms_text_message type queue using AQ?

    yes, i understand how to describe something...
    my question is how to get a list of properties.
    this data type has get functions that take the name of a property. now how do i know that property name to get it?

  • HTML Editor size limitation?

    Hi,<br>
    I'm using Apex 3.1 on Oracle 10g...<br>
    Is there a size limitation on the number of characters in the HTML Editor? <br>
    We have a Report/Form combo for a column. clicking on the "Edit" Link from the Report takes you to the Form where you can edit it. <br>
    The corresponding database column is of type LONG. I see the data in the report, but clicking on Edit gives me an error - <br>
    ORA-20001: Error fetching column value: ORA-06502: PL/SQL: numeric or value error: character string buffer too small <br>
    I've tried switching from CLOB to LONG for the column's datatype, but both fail to open the Form. <br>
    If I change the Item type to "Text Area", the LONG column works, but I do need it as an HTML Editor... <br>
    Thanks for any help!

    Scott (sspafado),
    Before you ask, my name is Viji
    :)

  • Size limitation for system-jazn.xml

    Hi helping hands,
    In sytem-jazn-data.xml and users-properties.xml , I want to include 1 lakh users. When I tried 1 lakh users,the bpel engine is not getting restarted.
    However I tried with 50k list of users, it works very fine.
    Is there any file size limitations are there for those 2 files.
    Could you suggest me an idea.
    Regards
    Viki

    Hi,
    I had to lookup what 1 lakh means and it appears to be 100.000? You should not put that many users in the system-jazn.xml file, not even 1000. It results in poor performance. You definitly should switch to OID or any other third party ldap server if you have that many users.
    Kind Regards,
    Andre

  • Max heap size limits

    I've been looking around for information on the max heap size limits on Sun's JVMs but can't seem to find any information. Just by testing, it seems like the max heap size for Windows 2000 can vary from 1.3G to 1.6G depending upon the machine (JDK 1.4). Does anybody know where I could find actual documentation that describes the limits for Sun's VMs on Windows (2000 and Advanced Server), Linux, and Solaris? I'm about to file this as a documentation bug against the JDK.
    God bless,
    -Toby Reyelts

    There was an older thread in the forums that had some info on this - my quick search failed to locate it, ypu might want to spend some time looking. The basic problem is memory space fragmentation by the OS, where the OS locates items in memory and effectively constrains heap growth to the unfragmented area that the heap starts in.While there may be more "unused" memory, it's not contiguous. There is also some info in MS's MSDN data regarding this condition, with information on the various OS's. I think Linux has a similar "condition".

Maybe you are looking for

  • Error while executing query on portal-SAPMSSY1 method : UNCAUGHT_EXCEPTION

    Hello all, I have created a quey and executed it on portal and everything just works fine. The query has few input varaibles at the start like fiscal year and some other inputs. I saved the query in my Portal favorites and then when I try to excute i

  • TREE in BSP

    Hi all, I want to display  a tree  with drag and drop  behaviour in BSP page. Could anyone help me how to proceed with this? Regards, keerthi

  • Update Key and Primary Key

    Hi When I am trying to define Update Key which is the other name for Alternate Key, then the primary key symbol (key) which is defined just beside my primary key column ( customer id) gets dissappeared and the new key symbol comes up besides the upda

  • Hidden Folders showing up in Color directory browser

    Are the hidden folders supposed to show up when i browse my directories from within Color? is this the way the program is supposed to work? Anyone know is this is a bug and if it is will there be a fix for it soon? thank you in advance, ~X

  • Anyone know "what version" of Word that Pages exports to?

    I have a business need to know what version of Word Pages exports as? It doesn't state anywhere and I am assuming if it is not a .docx it must be an older version. Any clue or ideas?