Data sent from backend exceeds buffer size

can some body please tell what is this error Data sent from backend exceeds buffer size

Without a context when this happens and especially where it´s impossible.
Markus

Similar Messages

  • How is goods receipt and invoice sent from backend system to SRM system?

    Hi friends
    How is goods receipt and invoice sent from backend system to SRM system?
    Thanks
    regards
    Krishna

    Hi Khan
    BBP_GET_STATUS_2 and Clean_recrec_Up are background jobs to sync ecc and ebp.
    BBPIV and MBGMCR are used to send goods receipt and invoice from srm to backend system
    So we dont require any special idocs to send goods receipt and invoice from ecc to ebp.Are all handled by BBP_Get_status_2
    Regards
    Krishna

  • Unexpected database connector error, when data sent from POJO.. Plz Help Me

    Hi, Please Help me about this, I have started learning to develop a crystal report with java swing in crystal report integrated eclipse IDE 3.5 but when i am trying to send data from pojo to report file its giving error like "Unexpected database connector error" and Error code:-2147215349 Error code name:docNotReady.... please hlep me... i uploaded my program and the link to download my file is :
    http://www.4shared.com/file/mz8O4iQi/SimplePojoReport.html
    and i am not using any database just setting the data in the pojo...
    i wrote my problem here before, but nobody replied me... so this time i uploaded my project.. please help me...
    thanks...

    Hell Ted Ueda,
          Thanks for your reply, and did you downloaded my project files ? if not then please download it first.. because i retrieving data from the POJO not from any database.. from database its working fine but in case of POJO its not, because i wanna retrieve data dynamically from directly fields to report, means that user will fill all the information in the text box (Swing) and when he click on the button then the data from the text box should be pushed to report file and also should have to saved in the database.. so please download my project file... and help me...
    link : http://www.4shared.com/file/mz8O4iQi/SimplePojoReport.html
    thanking you..
    Arun...

  • Performance affected from the socket buffer size

    I have a server program deployed in the Tomcat server.
    I found the server was sometime blocked in the native method java.net.SocketOutputStream.socketWrite0(Native Method). And the CPU was not fully utilized.
    Intuitive thinking, may be I can solve this problem by increasing the socket send buffer size. So I tried it.
    By default, tomcat socket send buffer size is 9000 bytes or so. I increased this value to 102400 bytes. Tested again. CPU usage was around 100%. OK, it worked.
    But how about if I decrease this value to a small number? I tried. Set the value to 100 bytes. Tested again. CPU usage was still around 100% !!!
    So, the problem here is: CPU was not able to be fully utilized by using the default buffer size (9000 bytes). But if you increase the value to a very large number or a very small number, you can achieve a better performance.
    note: the client was sending requests all the time. Just like the stress test. So, the server side was always busy.
    Edited by: willpowerforever on Jul 31, 2009 6:37 PM

    Thanks for all your reply.
    I give u more detail on my issue:
    Test Environment:
    Server: 4 GB memory, 1.3 GHz, a dual core CPU with hyper-threading enabled on each core (4 logical CPUs), a tomcat with my application deployed
    Client: the same configuration as Server, a http simulator
    Network: 1Gb/s
    Test:
    1> Simulate one user. Client keep sending requests to server, and server has one thread serve user's requests. (make Server always busy)
    request
    Server <------------- Client
    response
    Server ---------------> Client
    request
    Server <-------------- Client
    CPU usage was 18%, in normal case should be 25% (100% / 4 = 25%, in my other cases, CPU were 25%)
    Take a few thread dumps: saw threads sometimes blocked in the native method java.net.SocketOutputStream.socketWrite0(Native Method).
    2> Simulate 4 concurrent users. Client keep sending requests to server concurrently, and server has multiple threads serve user's requests. (make Server always busy)
    4 requests
    Server <------------- Client
    responses
    Server ---------------> Client
    requests
    Server <-------------- Client
    CPU usage was 80%, in normal case should be 90%+.
    Take a few thread dumps: saw threads sometimes blocked in the native method java.net.SocketOutputStream.socketWrite0(Native Method).
    If CPU was not fully utilized, for example only 80%, server must be blocked by something. And this bottleneck would impact the scalability of server.
    If I can remove this bottleneck and do not bring other overheads, I may acquire:
    1> a higher throughput
    2> a more scalable server
    In our performance requirement, server CPU usage must be 90%+.
    I just feel strange that why the CPU usage was normal (25% for single thread test or 90%+ for 4 threads test) even I choose a very small buffer size.

  • Find data sent from CRM to XI

    Hi ,
       We are sending some data from CRM to XI system.Is there any way or transaction code by which we can check that on particular date what data has been sent by CRM system to XI.
    its an argent requirement so if any one have any idea , it would be a great help for us.
    No need to mention that points will be awarded.
    regards
    Pankaj Giri

    please go to the tcode SXMB_MONI and check the message

  • Unwrap not unwraping all  data sent from server

    java client server NIO ssl
    At server end send about 50 packets are written using this method:
    sendPacket(String packet)
    writeBuffer = charSet.encode(packet);
    c.ssl.outNetBB.clear();
    c.ssl.engine.wrap(writeBuffer, c.ssl.outNetBB);
    c.ssl.outNetBB.flip();
    //write message to socket channel
    while (c.ssl.outNetBB.hasRemaining())
    socketChannel.write(c.ssl.outNetBB);
    At the client end I read all these packets:
    inNetBB.clear();
    int bytesread = socketChannel.read(inNetBB);
    inNetBB.flip();
    requestBB.clear();
    res = engine.unwrap(inNetBB, requestBB);
    It seems like all the 50 packets are read into inNetBB 6000 bytes
    But the unwrap only consumes 70 bytes and produces 100
    This first 100 bytes is unencrypted correctly and corresponds exactly to one write of the 50 from the server. I dont know why.
    Why am I getting only the first packet of the 50 sent??
    [  Ive tried : while inNtBB.hasRemaining for repeated unwraps
    but only the first packet is output still ]
    TIA
    pbutler

    If I'm reading this right, you're writing (wrapping) about 50 separate SSL packets, and then unwrapping just one.
    Recall:
    The SSLEngine produces/consumes complete SSL/TLS packets only, and
    does not store application data internally between calls to wrap()/unwrap().So it looks to me like you need to keep unwrapping the rest of your 5930 bytes to get the remainder of the application data.
    [ Ive tried : while inNtBB.hasRemaining for repeated unwraps
    but only the first packet is output still ]Are you inadvertantly clearing the inNtBB before you call your additional unwraps?
    If you have lots of small packets, you might want to consider buffering them up and wrapping them to the SSLEngine/SocketChannel all at once. You'll pay less in overhead.

  • How to display data sent from a midlet to a servlet over an outputstream in

    a browser??
    Hello
    Currently I send data to a servlet from a midlet and the servlet recieves and processes the data (confirmed through breakpoints) but I want to display the data in a JSP or directly thru the servlet using html tags.
    However I can forward or redirect the data from the servlet to my jsp code but this still doesn't launch a browser.
    A little more detail... I am not sending information from a JSP to the servlet. The servlet is listening on a port for input, so the servlet is being invoked directly when the data comes in. In the current state my servlet sends the results in a session object to a JSP which gets the results but can't display them since it doesn't launch browser.
    Does anybody know how I can achieve what I am trying to do or maybe suggest a better way to do it?
    Thanks

    The value of a list box will either be a string (if none or one item is selected) or an array (if more than one item is selected). So your code has to determine which it is, and if an array (object), loop through it and get the text of each selected item and add the comma and space. If it's not an array, just use the returned value. Here's a simple custom Calculate script that you can use in the text field:
    // Custom Calculate script for text field
    (function () {
        // Initialize the string
        var s = "";
        // Get the value of the list box (string or array)
        var v = getField("your_listbox").value;
        // If more than one item is selected, the type of the value will be "object" (an array really)
        if (typeof v === "object") {
            // Loop through the array to build up a string
            for (var i = 0; i < v.length; i++) {
                if (s) s += ", ";  // Add a comma and a space if needed
                s += v[i];  // Add the selected value
            // Set this field value to the string
            event.value = s;
        } else {
            // Set this field value to the one item selected
            event.value = v;
    Replace "your_listbox" with the actual name of the listbox you're using.

  • Create Back ground job for data recieved from a legacy system

    Hi All,
    I have a requirement where i need to schedule a back ground job once a proxy is triggered for the address data sent from a Legacy system. This proxy is triggerred from SAP PI, then i need to create a job shedule where address data is created as BP's(Business Partners) and assigned to a already created Target Group. Then once the Target group is completed with all BP needs to assigned to aCampaign Id. As the data is large it takes lot of time to execute. Please Advice.
    Regards,
    Nagesh Thanneeru

    Hi
    If you can divide the data into two internal tables and later collect into the one internal table, this could solve yuor problem.
    Also you need to make sure that enough memory is available in the system. Check DB6cockpit for that.
    Also there are a lot of correction notes available with SAP for the error"TSV_TNEW_BLOCKS_NO_ROLL_MEMORY"
    Check what the job exactly does, and search for the relevant notes, as there are different notes for different application areas...
    if its a loading job(Infopackage) you can reduce the data packet size of the IP, but that might hamper the loading performance.
    Hope this helps
    Regards
    Shilpa

  • Fill a table with data coming from an RFC

    Hello everyone:
    I've followed the Weblog "How many lines of java code did i write for a simple Web Dynpro?"
    /people/durairaj.athavanraja/blog/2004/10/17/how-many-lines-of-java-code-did-i-write-for-a-simple-web-dynpro
    I've called an RFC and created a table with data coming from it (which is also a table). My question is, if in this table there's a field named "UserType" there are two possible values for this field:
    "userA"
    "userB"
    How can I get the table only show me the "userA" registers? The RFC does return all of the users, but when filling the table, can I put an if-else somewhere on my code?
    Thanks a lot
    Alejandro

    Hi Alejandro,
    Referring to the link provided "The logic of the filter process is not implemented in Web Dynpro. The application developer must implement the action to be executed."
    We would have to implement the action onFilter in the controller implementation. Ideally, we fill the data retrieved from backend into a List (java.util.List) (this could be done on init of view) and then subset the list after meeting the criteria in the action handler(say
    onActionFilterData(com.sap.tc.webdynpro.progmodel.api.IWDCustomEvent wdEvent).
    Having done this, you may bind the output list back to the node (shown in table)
    Regards,
    Chaitanya

  • Not receiving information sent from specific server

    I'm not receiving data sent from a specifc address. Other (non Verizon) users are able to get data from the address. I can send data to it and pinkg and tracert times are good. The vendor needs a check of verion servers to see if there's a specific problem related t otransmitting data from the address. Online Verizon help was of no use since they cant' understand that a ping is only for seeing if you can connect to the servere.
    The specific proble is addressed in the following from the data provider's internet support
     Layer 4(transport layer) traffic is not reaching you from our app server with the address https://os-bo-app04-04.boldchat.com/. They want to know if there's any Verizon issue related to this.

    That's a curious question response from a company. Verizon and any isp are only providing layer 1,2 & 3 of the osi model. Any would verizon be involved in layer 4 issues for a particular website? I would think that answer would be pretty obvious.

  • Exceeds data buffer size discarding this snmp request

    Morning
    Cisco Prime LMS 4.2.3 is sending SNMP request too big for asa interface buffer.
    LMS is running on Windows server
    incoming SNMP request (528 bytes) from IP address x.x.x.x  Port  50592  Interface "inside" exceeds data buffer size, discarding this SNMP  request.
    212005: incoming SNMP request (%d bytes) from %s exceeds data buffer size, discarding this SNMP request.
    It is very much like this error
    Error Message    %PIX-3-212005: incoming SNMP request (number bytes) on interface
    interface_name exceeds data buffer size, discarding this SNMP request.
    Explanation    This is an SNMP message. This message reports that the length of the incoming SNMP  request, destined for the firewall, exceeds the size of the internal data buffer (512 bytes) used for  storing the request during internal processing; therefore, the firewall is unable to process this request.  This does not affect the SNMP traffic passing through the firewall via any interface.
    Recommended Action    Have the SNMP management station resend the request with a shorter length,  for example, instead of querying multiple MIB variables in one request, try querying only one MIB  variable in a request. This may involve modifying the configuration of the SNMP manager software.
    how do I change the SNMP request size in LMS?
    I can only find the following that might be an option
    http://blogs.technet.com/b/mihai/archive/2012/05/14/reducing-the-maximum-number-of-oids-queried-in-a-single-batch-in-om-2012.aspx
    any thoughts on the matter would be appreciated
    just using default settings with snmpv3

    Bug in lms 4.2.3
    CSCtj88629            Bug Details
    SNMP packet size requests from LMS is too large
    Symptom:
    LMS sends more than 512 SNMP requests to the FWSM, so it rejects the requests.
    Conditions:
    This occurs with FWSM and ASA's.
    Workaround:
    None.
    http://tools.cisco.com/Support/BugToolKit/search/getBugDetails.do?method=fetchBugDetails&bugId=CSCtj88629

  • Why replay history data from NI-DAQ buffer?

    LabVIEW 8.6.1f1, NI-DAQ 8.8, Win Vista, two cDAQ chassis.  (I can't move to LV09 or LV10 for this, not yet anyway).
    I have a client with two cDAQ banks, each with 16 TC channels (on 4 modules) and 16 voltage channels (on 1 module), 64 channels all told.
    The config operation is like this:
    This is called twice, with TC CHANNELS and V CHANNELS describing the channel arrangement.  Both channel descriptors are the same except for device number. The Sample rate works out to 1000 / 100 = 10 Hz. I am using CONTINUOUS SAMPLES, with no need for a buffer, but I understand that one is created anyway.
    One loop of my main program is waiting on the READ function:
    You can see the explanation of why the loops are there.
    After reading the data (at 10 Hz), each frame is sent via UDP to somewhere, and every 5th frame is displayed, via another loop.
    This program runs 24/7 on the same computer as another program.
    All this works fine in normal circumstances.
    However, occasionally, a weirdness creeps in:
    In the mornings, after waking the screensaver (or the CPU itself, I'm not sure), we sometimes (not every time) get a very rapid display - the displays are changing much faster than the expected 2 Hz.
    I've been wondering whether the data shown was historical, or simply faster updates of live data.
    Today, it so happened that a blower was turned on just before waking the computer. The device takes 15 seconds to ramp from zero up to operating pressure normally, and then it stays there.
    The readings on screen were frozen (unchanging).  When the computer was awakened, what my client saw was the reading for the blower pressure ramp from zero to running pressure in 2-3 seconds, stay there for 2-3 seconds, drop quickly to zero for 4-5 seconds, and start the ramp again.
    Over and over and over it replayed this data, way faster than 2 Hz.  At least 40 times it replayed this same data, while the real pressure was just variations around the operating pressure.  As time went on, instead of dropping to zero, it would drop to some point on the ramp up and repeat from there.  
    I don't know if it would eventually straighten itself out - the client gives up and restarts the program.  All is well after that.
    I've gone over my code several times, and I can't see a mechanism for me to store this data and re-play it.
    Here's my best guess as to what's happening:
    Since I am not explicitly setting the buffer size, it defaults to 1000 samples.
    Since I'm running at 10 Hz, 1000 samples = 100 sec.
    At some point, because of the screensaver, or because of CPU sleepage, or because of integer rollover, or because of zombie infiltration during the night, NI-DAQ has become confused. 
    When the mouse was wiggled, the gates were opened. After that, whenever I ask for a sample, it gives me one, without regard to read/write position. It continually gives me the next sample from the buffer and wraps around, again and again, and again, as fast as I can read / process / display.
    It is also putting new data into the buffer, so that the historical data eventually gets wiped out.
    That fits the facts:
      It shows only recent (last 100 sec) data, not ancient data, and not live data too fast, as I had considered.
      The data changes with time - oldest data gets wiped out.
      It plays the SAME data over and over and over.
    But that means a bug in NI-DAQ, I think.  
    QUESTIONS:
    Any flaws in my thinking?
    Any previous sightings of this symptom?
    Any ideas for workarounds?
    Steve Bird
    Culverson Software - Elegant software that is a pleasure to use.
    Culverson.com
    Blog for (mostly LabVIEW) programmers: Tips And Tricks

    The rapidly changing data, is that on the viewing machine or on a display on the device acquiring the data?
    The "device acquiring the data" is a pair of cDAQ chassis.  There is no display there. It's the LabVIEW program that is reading and displaying the data.
    Does your computer go into any sort of power saving mode where the USB controllers are disabled during hibernation/standby?
    on MY machine, I have tried this with CPU sleepage and such.  Didn't know you could sleep the USB, but on MY machine, that's disabled.
    But I haven't made it happen on MY machine.
    My CLIENT's machine is actually running XP, and is set to turn off monitor NEVER, turn off hard disk NEVER, and system standby NEVER.
    What version of DAQmx do you have installed?
    On MY machine, it's 8.8 - I haven't found out from my client yet what his is, but it won't be far away from what came with LV 8.6.
    Can you post your code?
    Well, OK:
    Here's two of the loops running (the third loop deals with TCP/UDP and is not relevant, I think).
    The TCM DAQ manager 's READ function is pictured in the original post.
    Here we wait on a sample, scale it (volts into EU), and feed it to the PROCESS TCM DATA vi (which does averaging).
    Here's the NEW DATA function for PROCESS TCM DATA:
    The ST (short-term) HISTORY is set up as an N channel by 5-sample array, and ST HISTORY SAMPLES is set to 5.
    The DAQ loop, after processing this data, triggers a DISPLAY event, which triggers a DISPLAY action, which lands here:
    The GET DISPLAY function retrieves the latest averages:
    That data (64 channels worth) is split into two banks and displayed:
    The display process simply distributes the array data into clusters (channels can be user-configured to be in various locations on the screen):
    The UPDATE TIP action causes a TIP STRIP string to be updated:
    The TIP STRIP shows the average and min and max values (which might be hidden from the display itself), based on which display cell the mouse is in.
    A MOUSE ENTER event on any display cell sets the CLUSTER CONTAINING MOUSE variable to the CTL REF of the event, a MOUSE LEAVE event for the overall display bank (containing 32 cells) sets the CLUSTER CONTAINING MOUSE variable to Not-A-RefNum.
    the DISPLAY looks like this (only a few cells shown).  The display for this program occupies an entire 1680x1050 LCD display, another program runs on the main monitor.
    As you can see, the only storage in this process is the AVERAGER, and that is limited to 5 samples (0.5 sec). It cannot store any more than that.
    I've thought about the possibility of the mouse oscillating between two cells (as mice are sometimes known to do) when it is parked, and causing a flood of MOUSE ENTER events.
    But 1) it happens more often than the odds would dictate  (a random location is not likely to be on a cell border), and
    2)  When that cleared up, the queue would be processed, but it would display CURRENT data more rapidly than 2 Hz. NOT historical data.
    Given that it seems to be about 100 sec of history (judging by my client's recap of what he saw, and how it operates), that points to the 1000-sample buffer within NI-DAQ.  That's the only place that I see to store that data.
    I'll admit that I'm not watching the ERROR output from the NIDAQ READ function. Perhaps I should.
    The only scenario I can come up with is that NI-DAQ's READ pointer is confused.
    When I ask for a sample it tells me that there is ONE and only ONE available.  So I deal with it.
    When I ask again, it tells me the same thing.
    When I ask again, it tells me the same thing.
    It goes round and round the READ buffer.  At the same time, new data is coming in, replacing the oldest stuff.
    But it (apparently) never gets back to the "normal" condition, where it WAITS on a sample at the RAD function.
    Nobody has waited long enough to see if it would do that eventually, after a few minutes (50+ replays), they give up and restart.
    Restarting the program resets the cDAQ drivers, among other things, so the problem goes away.
    If there's another scenario to cause these symptoms, I'm all ears. But I cannot think of it.
    If it was a USB SLEEP problem, I wouldn't think the data would replay from BEFORE the wake-up. But that's what happened.
    Steve Bird
    Culverson Software - Elegant software that is a pleasure to use.
    Culverson.com
    Blog for (mostly LabVIEW) programmers: Tips And Tricks

  • Linux Serial NI-VISA - Can the buffer size be changed from 4096?

    I am communicating with a serial device on Linux, using LV 7.0 and NI-VISA. About a year and a half ago I had asked customer support if it was possible to change the buffer size for serial communication. At that time I was using NI-VISA 3.0. In my program the VISA function for setting the buffer size would send back an error of 1073676424, and the buffer would always remain at 4096, no matter what value was input into the buffer size control. The answer to this problem was that the error code was just a warning, letting you know that you could not change the buffer size on a Linux machine, and 4096 bytes was the pre-set buffer size (unchangeable). According to the person who was helping me: "The reason that it doesn't work on those platforms (Linux, Solaris, Mac OSX) is that is it simply unavailable in the POSIX serial API that VISA uses on these operating systems."
    Now I have upgraded to NI-VISA 3.4 and I am asking the same question. I notice that an error code is no longer sent when I input different values for the buffer size. However, in my program, the bytes returned from the device max out at 4096, no matter what value I input into the buffer size control. So, has VISA changed, and it is now possible to change the buffer size, but I am setting it up wrong? Or, have the error codes changed, but it is still not possible to change the buffer size on a Linux machine with NI-VISA?
    Thanks,
    Sam

    The buffer size still can't be set, but it seems that we are no longer returning the warning. We'll see if we can get the warning back for the next version of VISA.
    Thanks,
    Josh

  • Sychronize AO/AI buffered data graph and measure data more than buffer size

    I am trying to measure the response time (around 1ms) of the pressure drop indicated by AI channel 0 when the AO channel 0 gives a negetive single pulse to the unit under test (valve). DAQ board: Keithley KPCI-3108, LabView Version: 6.1, OS system: Win2000 professional.
    My problem is I am getting different timed graph between the AI and AO channels every time I run my program, except the first time I can get real time graph. I tried to decrease the buffer size less than the max buffer size of the DAQ board (2048 samples), but it still does unreal time graph from AI channel, seems it was still reading from old data in the buffer when AO writes the new buffer data, that is my guessing. In my p
    rogram, the AO and AI part is seperated, AO Write buffer is in a while loop while AI read is not. Would that be a problem? Or it's something else?
    Also, I am trying to measure data much larger than the board buffer size limit. Is it possible to make the measurement by modifying the program?
    I really appreciate any of your help. Thank you very much!
    Best,
    Jenna

    Jenna,
    You can modify the X-axis of a chart/graph in LabVIEW to display real-time. I have included a link below to an example program that illustrates how to do this.
    If you are doing a finite, buffered acquisition make sure that you are always reading everything from the buffer for each run. In other words, If you set a buffer size of 5000, then make sure you are reading 5000 scans (set number of scans to read to 5000). This will assure you are reading new data every time you run you program. You could always put the AI Read VI within a loop and read a smaller number from the buffer until the buffer is empty (monitor the scan backlog output of the AI Read VI to see how many scans are left in the buffer).
    You can set a buffer size larger than the FIFO
    buffer of the hardware. The buffer size you set in LabVIEW is actually a software buffer size within your computer's memory. The data is acquired with the hardware, stored temporarily within the on-board FIFO, transferred to the software buffer, and then read in LabVIEW.
    Are you trying to create a TTL square wave with the analog output of the DAQ device? If so, the DAQ device has counters that can generate a highly accurate digital pulse as well. Just a suggestion. LabVIEW has a variety of shipping examples that are geared toward using counters (find examples>>DAQ>>counters). I hope this helps.
    Real-Time Chart Example
    http://venus.ni.com/stage/we/niepd_web_display.DISPLAY_EPD4?p_guid=B45EACE3E95556A4E034080020E74861&p_node=DZ52038&p_submitted=N&p_rank=&p_answer=&p_source=Internal
    Regards,
    Todd D.
    National Instruments
    Applications Engineer

  • Data retrieval buffers - buffer size and sort buffer size

    Any difference to tune BSO and ASO on data retrieval buffers?
    From Oracle documentation, the buffer size setting is per database per Essbase user i.e. more physical memory will be used if there are lots of concurrent data access from users.
    However even for 100 concurrent users, default buffer size of 10KB (BSO) or 20KB (ASO) seems very small compare to other cache setting (total buffer cache is 100*20 = 2MB). Should we increase the value to 1000KB to improve the data retrieval performance by users? The improvement impact is the same for online application e.g. Hyperion Planning and reporting application e.g. Financial Reporting?
    Assume 3 Essbase plan types with 100 concurrent access:
    PLAN1 - 1000KB * 100 = 100MB 1000KB * 100 = 100MB (total sort buffer size)
    PLAN2 - 1000KB * 100 = 100MB 1000KB * 100 = 100MB (total sort buffer size)
    PLAN3 - 1000KB * 100 = 100MB 1000KB * 100 = 100MB (total sort buffer size)
    Total physical memory required is 600MB.
    Thanks in advance!

    256 samples Buffer size will always give you a noticable amount of latency. If you use Software Monitoring you should try setting your Buffer to 64 samples. With the recording delay slider in the preferences->Audio you can compensate for the latency (of course not in realtime) so that the Audio will be placed exactly there, where it should hae been recorded at. In your case set it to a -value. A loopback test (check the link below) will clarify the exact amount of Latency occuring on your system.
    http://discussions.apple.com/thread.jspa?threadID=1883662&tstart=0

Maybe you are looking for

  • Cannot write or delete files from DVD+RW after erasing it through Disk Utility. Help.

    Hello. I erased (completely not quickly) my 4.7 GB DVD+RW with the help of Disk Utility in which it was successful, it asked me to open Finder, DVD Player or Disk Utility, I chose Finder. Then I dragged a MP4 and burned it, it asked me to name the bl

  • IE7 issue with website

    I developed my own website in Dreamweaver CS4 checking browser compatibility and validated all pages for both IE and Firefox. I uploaded my files to my Yahoo hosting account and now that it is live I can not view my home page (index.html) and one oth

  • Error handling in lsmw

    Hi, Is it possible to handle errors in LSMW if so how? Advantages of LSMW over Function Module. Is it possible to create multiple structures in LSMW to accept multiple Flat files. Thanks in Advance.

  • Dreamweaver Template alignment Issue

    I am using Dreamweaver 8, i have created a template with my background image, my menu, and an outter table to align my content. I have entered new content in one of the pages that I have applied the template to, and now all of it is shifted to the ri

  • Database polling process

    I have created a bpel database polling process to process a single table and use a logical update (mark a column as processed). The table will contain many unprocessed records at any time. I have setup the polling interval 5 Sec and records unlimited