Buffer Qua is 60 % , incre the sga_max_size and db_cache_size on wht basis

Dear All,
                    In our XI Production system, data buffer quality is 60%, as i found out that sga_max_size(1.1 GB) and db_cache_size(570 MB) not set as per the standard, and also they are less when compared with development(8 GB RAM).
                                             Our RAM size in PRD is 10 GB,and when i see in topas 90% of RAM is filled, if we increase these parameter by 1GB, i think System's overall  perform will be effected as the total RAM will be filled. Please suggest me how to solve the problem and on what basis these parameters should be set(what percent of RAM).
OS : AIX 5.3
SAP XI 7.00
DB : Oracle
Please inform whether i have to include any other info
Regards,
Balaji Vedagiri

a) Why was it changed to less than default? Was the purpose to give memory to Java stacks?
Ans: These parameters are set long back
tell me how much RAM memory does the ABAP+JAVA stack occupy(minimum), as 90% of  RAM  is filled now.
b) Is there a performance issue now, XI is not a database heavy applicaton? If not why bother?
Ans:  Currently there are not performance issues, but in EWA report it is RED and my customer wants it to increase the quality to atleast by 90%.
c) Why not change it back to default or to the same as QA?
Ans: As my analysis we need to increae the parameters by 1GB both means, the remaining RAM(1GB) will be filled means, my total RAM will filled this will lead to overall performance problem.
Regards
Balaji Vedagiri

Similar Messages

  • Increase the SGA_MAX_SIZE and SGA_TARGET parameters

    HI ,
    I am using oracle 10g R2 and windows server 2003 .
    My server is having 8 GB RAM .
    I want to increase the SGA_MAX_SIZE and SGA_TARGET parameters .
    I am using the below command , but getting the error
    SQL> alter system set sga_max_size=2000M scope=spfile ;
    alter system set sga_max_size=2000M scope=spfile
    ERROR at line 1:
    ORA-32001: write to SPFILE requested but no SPFILE specified at startup
    Can someone please help ..

    hi
    connect as sys user and then try...
    first make folder as backup and then copy the old SPFILEorcl to the same folder(backup)
    now you can easily modify the files otherwise you will get error..
    you will find here..
    C:\ORACLE_HOME\database
    SQL>  alter system set sga_max_size=2000M scope=spfile ;
    System altered.OR.
    connect as sys user.
    and do the following...
    1).SQL>create pfile from spfile;
    2).modify the pfile and decrease\increase the size of pools.
    3).shutdown database.
    4). Startup database using created pfile.
    if it start ok then.
    5.) Recreate the Spfile with the created Pfile.
    SQL>create spfile from pfile;sarah

  • What are the Roles and Responsibilities as a Basis Consultant in SRM

    Hi Gurus,
             I want to know what are the activity in SRM as a Basis Consultant. I know SRM installation,, but i have no idea how to monitor  and configure SRM server ,, pls help me,,
    regards,
    Balaram

    HI balaram reddy 
    You can use the administration guide under the below link.....
    service.sap.com/instguidesu2192 mySAP Business Suite Solutions u2192 mySAP SRM u2192 Using SAP SRM Server 5.5 u2192Solution Operations Guide: mySAP Supplier Relationship Management
    Please find the links as requested. You will find lot of documents and so on here.
    http://service.sap.com/srm
    http://help.sap.com/saphelp_srm50/helpdata/en/0e/cede3b6bc8682be10000000a11402f/frameset.htm
    Regards
    Bandla

  • SD invoice not showing the Ecess and SEcess

    Hi SD consultants,
    I am executing a depot sale scenario.
    I craeted a SO, Delivery and billing in VF01. When i check the billing doc ina ccounting view it is not picking the ECess and SHEcess.
    Where as in SO i can see the ECess and SHEcss .
    Wht could the wrong.
    Also after invoice creation i am creating excise invoice in J1Ij and refering the factory invoice . I need to know is setoff will happen ( Monthly utilisation in depot level because we r charing ED from customer )
    Request all to reply
    Thanks
    A.Gururajan

    Hi
    Kindly check the belwo
    1. Check the coditio n records maintained for the condition types and the validity periods
    2. Ch3eck the pricing procedure in the billing document
    Regards
    Damu

  • My mouse/trackpad are completely out of control. The cursor/arrow is flying all over the screen and closing/opening windows. What is the solution to this? Downloading windows for Mac?

    So far I'm dissapointed with Mac O/S .....so many areas lack common sense, require unnessary extra clicks and is not as customizable as Windows. It's like Apple thinks you are a child and not entitled to choose for yourself so that they can control your info for you.
    The mouse/trackpad problem is a major issue that Apple should be proactive in offering a solution for. If anyone has a solution to this I'd be greateful to hear it.
    Thanks, 

    Is this the first time you're using a Mac?
    If so the learning curve is somewhat steep - but, if you put in the time and use this knowledge base whenever you run into issues you'll be up to speed soon enough.
    With regards to the trackpad/mouse, you can adjust the sensitivity by:
    1. Click the Apple Logo on the top left hard corner of your screen
    2. Click System Preferences
    3. Click Trackpad
    Adjust the tracking speed to your specifications.
    Also familiarise yourself with the gestures (activate the ones you can remember)
    To go back to the menue by clicking Show All
    4. Click Mouse
    Adjust your mouse settings to your specification.
    And why use mac?
    It just works...
    (most of the time ;p)

  • I need to download an online quicktime video for a project.  Its is 3gb so takes a while to buffer but when it buffers to about half way it stops and I have to reload the page and try again?

    I need to download an online quicktime video for a project.  Its is 3gb so takes a while to buffer but when it buffers to about half way it stops and I have to reload the page and try again?

    Let's first try updating using an installer file downloaded from the Apple Website:
    http://www.apple.com/itunes/download/

  • Refresh buffer when rfc gets data from crm and the data was updated from R3

    hi all
    I have a program in r3 for maintenance of partners, so the user can create or update a CRM partner from the R3 system. if you are updating a partner, the program gets the data from CRM via RFC and saves the changes via zBDoc. if the user modify a partner and then try to check if the info were updated, he needs to leave the transaccion and run it again in order to the RFC gets the data updated. how I can avoid to leave the transaccion? i tried with a buffer refresh and commit but it doesn't works.... any idea?
    tnks

    Hi Alan,
    THANK YOU for that hint. That brought me a giant step further, seriously! Many thanks! Now my report is running and hourly creating an instance, reeally nice!
    But of course, the next problem is arising rapidly: Now that I got my report creating instances, the funny thing happening is the following: I created the WebI report with the WebI-Rich Client and then saved it to the server. In InfoView I scheduled the instance creation and the strange thing is now: The values differ. In the original report the values for some percentage ratios (e.g. 50%) show up as they should with four decimals: 0,5000 (German way of displaying numbers with decimals). BUT: In the instances of the same report, exactly these numbers would appear as 0,0000. So the numbers would just not be displayed! But all other numbers (all numbers not having decimals) would show up as they are supposed to even in the instances of the report.
    This is kind of strange to me. Does anybody have a clue what the problem might be?
    Many thanks again! And I hope I am not bothering you guys with these "newbie" questions.
    Thanks and best regards,
    Kai

  • Relative size of sga_max_size and sga_target to the physical memory

    Dear Experts,
    I have installed oracle 10.2.0.4 on AIX 5.3 with 32 GB of physical memory.
    Let us say other applications on the server consume 10 GB of physical memory.
    Now how should i plan the size of SGA_TARGET and SGA_MAX_SIZE on the same.

    As per Oracle® Database Performance Tuning Guide 10g Release 2 (10.2):
    You must then divide the resulting memory between the SGA and the PGA.
    For OLTP systems, the PGA memory typically accounts for a small fraction of the total memory available (for example, 20%), leaving 80% for the SGA.
    For DSS systems running large, memory-intensive queries, PGA memory can typically use up to 70% of that total (up to 2.2 GB in this example).
    Good initial values for the parameter PGA_AGGREGATE_TARGET might be:
    For OLTP: PGA_AGGREGATE_TARGET = (total_mem * 80%) * 20%
    For DSS: PGA_AGGREGATE_TARGET = (total_mem * 80%) * 50%
    where total_mem is the total amount of physical memory available on the system.
    (http://download-west.oracle.com/docs/cd/B19306_01/server.102/b14211/memory.htm#i49320)

  • How do i clear the receiver and send buffer?

    I am trying to communicate with the Votsch 3050 thermal chamber, with LabVIEW as the control software via serial RS232. In sending the required ASCII strings, there was no response from the chamber, yet no error.
    The manufacturers indicate clearing the receive and send buffer using the the "PurgeComm(HANDLE hFile, DWORD dwFlags)" function in C, but offer no LabVIEW support.
    Does any one have any ideas on how i can do this using LabVIEW? Any Information or VIs would be greatly appreciated.

     @dsbuxi I have worked with many patient monitors. They often work in a client server setup. Data is not sent before you request it. And the communication has to follow a certain protocol. Your job will be to implement this protocol in Labview.  This is not too hard, but it requires some reading to understand the protocol. I am not able to send you any code, but feel free to ask if you need some help regarding the protocol 
    And you may have to order or make a custom cable. But you will also find this information in the protocol description. Good luck
    Message Edited by t06afre on 01-07-2009 09:03 PM
    Besides which, my opinion is that Express VIs Carthage must be destroyed deleted
    (Sorry no Labview "brag list" so far)

  • SGA_MAX_SIZE and AMM

    Hi, all,
    I've been testing using AMM in our newly-upgraded 11gR2 databases. All databases (Prod and non-Prod) are running 11.2.0.3 on AIX 5.3 (64-bit).
    I noticed that despite me explicitly setting the following to 0:
    alter system set db_cache_size = 0 scope=spfile;
    alter system set log_buffer = 0 scope=spfile;
    alter system set pga_aggregate_target = 0 scope=spfile;
    alter system set sga_max_size = 0 scope=spfile;
    alter system set sga_target = 0 scope=spfile;
    alter system set shared_pool_size = 0 scope=spfile;
    alter system set shared_pool_reserved_size = 0 scope=spfile;
    And allowing Oracle 'full control' how it manages all parts of the memory (SGA and PGA)
    -- The 'big' database
    alter system set memory_max_target = 24G scope=spfile;
    alter system set memory_target = 24G scope=spfile;
    -- The 'small' database
    alter system set memory_max_target = 3G scope=spfile;
    alter system set memory_target = 3G scope=spfile;
    I bounce both databases and things seem to come up OK. The only exception is that sga_max_size seems to be set to a specific value:
    For the 'big' database, it's 16Gb (2/3rds of the total memory_target)
    For the 'small' database, it's 1536Mb (1/2 of the total memory_target)
    I don't really understand why this is happening. I've double-checked my SPFILE, I've queried v$spparameter and I've read MOS 443746.1
    +"If MEMORY_TARGET is set to a non-zero value:+
    +If neither is set, they will be auto-tuned without any minimum or default values. We will have a policy of distributing the total memory set by MEMORY_TARGET parameter in a fixed ratio to the the SGA and PGA during initialization. The policy is to give 60% to the SGA and 40% to the PGA at startup."+
    BUT I presume that sga_max_size works in the same way that it did when we weren't using AMM - it's the upper limit of the SGA. Not ideal, to be honest, as I can definitely see times when the SGA needs a LOT of buffer cache in Production (and, therefore, lots of SGA)
    I do see this in the alert.logs whenever I restart the system:
    Using parameter settings in server-side spfile /opt/oracle/product/11.2.0.3/db/dbs/spfilebigdb.ora
    System parameters with non-default values:
    processes                = 600
    sessions                 = 928
    timed_statistics         = TRUE
    resource_limit           = TRUE
    event                    = "10262 trace name context forever, level 160000"
    sga_max_size             = 16G
    So it's suggesting that it's querying the SPFILE and determining that sga_max_size should be set. But I've double-double-checked and that isn't the case.
    Now, I know I can get around this by explicitly setting sga_max_size at memory_target. But I was wondering if anyone else had seen this before too?
    Mark

    Another workaround is to remove the memory components entirely from the SPFILE (instead of having them set to 0, don't have them in the file at all). This seems to also set the sga_max_size to the memory_target - which is expected behaviour given the MOS note.

  • To find out the min and max memory been used by each parameter under SGA_MA

    Hi,
    Can any please tell me how to find out the min and max memory been used by each parameter under SGA_MAX and SGA_TARGET ? below is the db CRMS65T. If any such script is there please provide me
    SQL> select name from v$database;
    NAME
    CRMS65T
    SQL> show parameter sga
    NAME                                 TYPE        VALUE
    lock_sga                             boolean     FALSE
    pre_page_sga                         boolean     FALSE
    sga_max_size                         big integer 1000M
    sga_target                           big integer 1000MThanks in advance

    Can any please tell me how to find out the min and max memory been used by each parameter under SGA_MAX and SGA_TARGET ? below is the db CRMS65T. If any such script is there please provide meI guess your question is, each memory components of SGA? if so
    SQL> select * from v$sgainfo;
    NAME                                  BYTES RES
    Fixed SGA Size                      2088504 No
    Redo Buffers                       18882560 No
    Buffer Cache Size                 616562688 Yes
    Shared Pool Size                  301989888 Yes
    Large Pool Size                     4194304 Yes
    Java Pool Size                      4194304 Yes
    Streams Pool Size                         0 Yes
    Granule Size                        4194304 No
    Maximum SGA Size                  947912704 No
    Startup overhead in Shared Pool   125829120 No
    Free SGA Memory Available                 0Also check
    SQL> select COMPONENT,CURRENT_SIZE,MIN_SIZE,MAX_SIZE,USER_SPECIFIED_SIZE from v$sga_dynamic_components;
    shared pool                                                         301989888  301989888          0           209715200
    large pool                                                            4194304    4194304          0             4194304
    java pool                                                             4194304    4194304          0             4194304
    streams pool                                                                0          0          0                   0
    DEFAULT buffer cache                                                616562688  616562688          0           603979776
    KEEP buffer cache                                                           0          0          0                   0
    RECYCLE buffer cache                                                        0          0          0                   0
    DEFAULT 2K buffer cache                                                     0          0          0                   0
    DEFAULT 4K buffer cache                                                     0          0          0                   0
    DEFAULT 8K buffer cache                                                     0          0          0                   0
    DEFAULT 16K buffer cache                                                    0          0          0                   0
    DEFAULT 32K buffer cache                                                    0          0          0                   0
    ASM Buffer Cache                                                            0          0          0           603979776
    13 rows selected.
    SQL>Edited by: CKPT on Sep 19, 2011 8:55 AM

  • How do I keep a Window resize from holding the CPU and temporaily interrupting the Labview application?

    I am performing a data aquisition in one VI at 10000 Samples per second, then averaging 100 samples every 10 msec.  This runs in a loop and I monitor the time of the loop.  Another VI pulls the single point result of each 10 msec average and plots it to a graph.
    If another window on the Windows XP operating system is re-sized the LabVIEW VI performing the Acquisition suspends until the resizing is complete and I can watch the loop time go from 10 msec to whatever time the other window is being manipulated (100, 200 300 msec...).  The resulting graphical display will then show the next averaged plot point (this is a point to point drawing) without any of the 'real' analog activity that occurred during these +10 msec interruptions (the dataq VI pulls more than 100 samples to perform the averaging).
    Does anyone know how I can prevent the resizing from taking such a high priority on the CPU?  For various reasons I do not want to change my data acquisition scheme.

    chrisger  says: "if you want to pass data without loss you should use a method that allows buffering of data, e.g. queues. Take a look at the LV documentation (there are several solutions, but a named queue is the easiest way to go for such sort of applications)."
    Actually, one VI is running as the executable and the other is called dynamically.  The Dynamic VI uses a queuing vi common to both to access the data.  This part works well, it is the interrupt of the main vi perfoming the data aquisition that is the problem. 
    Because I am not using the full buffer of data, but only whatever portion is acquired and then averaged, the graphing vi essentially gets two (software timed) points and draws between them. 
    I realize software timing is bad, but I am kinda stuck with this scheme.  So, I want to minimize the interruption when a non-related window is re-sized.

  • SGA_MAX_SIZE  and  SGA_TARGET  in Oracle 10G

    We have the following H/W with Solaris 9
    SPARC-IV 16 CPUs
    32GB RAM
    128GB swap
    What is the optimal size of sga_target and sga_max_size that can be configured for a heavy batch processing oracle server. Nothing runs except oracle in the above server.
    We have currently 3GB SGA size , however our heavy queries are very slow.
    We have huge updates on tables with more than 200Millions rows.
    Do we need to have JAVA_POOL if the java stored procedure are NOT used.
    However we are connecting to this database using JDBC also.
    Moreover ASM is enabled in this server.
    You expert suggetions are always welcome.

    That's not a question that can be answered easily. It depends very much on where the time is going. Three main targets:
    Writing the log file - which is largely about use of disks, not memory.
    Reading data - lots of single block reads suggest looking at a bigger cache
    Reading and writing temp because of large sorts/hash joins - pga configuration
    One difficulty with TEMP is that the I/Os are often asynchronous, so a session can do a lot of TEMP I/O without recording much time - which means it can cause a problem elsewhere without suffering a problem itself.
    You didn't mention the pga_aggregate_size - which is an important consideration in 10g and can be more important than the sga sizing in DSS and batch-like processing. Between them, the SGA and PGA_aggregate_target should probably account for a very large fraction of your memory - but (since you are on Solaris) make sure you look into large memory pages and intimate shared memory, otherwise you may need to leave a lot of memory available for the O/S memory management mechanisms.
    As far as the java pool is concerned, check v$sgastat for pools which constantly show free memory, and reclaim any large free volumes.
    Bottom line - if you've got the memory in the box, you might as well use it in the best possible place; but at the moment we don't have enough information to tell us where it would be most useful.
    Regards
    Jonathan Lewis
    http://jonathanlewis.wordpress.com
    http://www.jlcomp.demon.co.uk

  • Synchronize input and output tasks to start at the same sample point [C++ NI_DAQmx Base]

    I'm trying to initiate the analog input and output streams to start reliably at the same sample. I've tried triggering the output from the start of the input using the following code [NI-DAQmx Base 2.1 under Mac OS X with an M-Series multifunction board]. It compiles and runs, but gives an error message at the call to "DAQmxBaseCfgDigEdgeStartTrig". Any suggestions about synchronized I/O on this platform?
    #include "NIDAQmxBase.h"
    #include
    #include
    #include
    #define DAQmxErrorCheck( functionCall ) { if ( DAQmxFailed( error=( functionCall ) ) ) { goto Error; } }
    int main( int argc, char *argv[] )
    // Task parameters
    int32 error = 0;
    TaskHandle inputTaskHandle = 0;
    TaskHandle outputTaskHandle = 0;
    char errorString[ 2048 ] = {'\0'};
    int32 i;
    time_t startTime;
    // input channel parameters
    char inputChannelList[] = "Dev1/ai0, Dev1/ai1";
    float64 inputVoltageRangeMinimum = -10.0;
    float64 inputVoltageRangeMaximum = 10.0;
    // output channel parameters
    char outputChannelList[] = "Dev1/ao0, Dev1/ao1";
    char outputTrigger[] = "Dev1/ai/StartTrigger";
    float64 outputVoltageRangeMinimum = -10.0;
    float64 outputVoltageRangeMaximum = 10.0;
    // Timing parameters
    char clockSource[] = "OnboardClock";
    uInt64 samplesPerChannel = 100000;
    float64 sampleRate = 10000.0;
    // Input data parameters
    static const uInt32 inputBufferSize = 100;
    int16 inputData[ inputBufferSize * 2 ];
    int32 pointsToRead = inputBufferSize;
    int32 pointsRead;
    float64 timeout = 10.0;
    int32 totalRead = 0;
    // Output data parameters
    static const uInt32 outputBufferSize = 1000;
    float64 outputData[ outputBufferSize * 2 ];
    int32 pointsToWrite = outputBufferSize;
    int32 pointsWritten;
    for( int i = 0; i < outputBufferSize; i++ )
    outputData[ 2 * i ] = 9.95 * sin( 2.0 * 3.14159 * i / outputBufferSize );
    outputData[ 2 * i + 1 ] = -9.95 * sin( 2.0 * 3.14159 * i / outputBufferSize );
    // ------------------- configure input task -----------------------
    DAQmxErrorCheck ( DAQmxBaseCreateTask( "", &inputTaskHandle ) );
    printf( "Created input task\n" );
    DAQmxErrorCheck ( DAQmxBaseCreateAIVoltageChan( inputTaskHandle, inputChannelList, "", DAQmx_Val_RSE, inputVoltageRangeMinimum, inputVoltageRangeMaximum, DAQmx_Val_Volts, NULL ) );
    printf( "Created AI Voltage Chan\n" );
    DAQmxErrorCheck ( DAQmxBaseCfgSampClkTiming( inputTaskHandle, clockSource, sampleRate, DAQmx_Val_Rising, DAQmx_Val_ContSamps, samplesPerChannel ) );
    printf( "Set sample rate\n" );
    // ------------------- configure output task -----------------------
    DAQmxErrorCheck ( DAQmxBaseCreateTask( "", &outputTaskHandle ) );
    printf( "Created output task\n" );
    DAQmxErrorCheck ( DAQmxBaseCreateAOVoltageChan( outputTaskHandle, outputChannelList, "", outputVoltageRangeMinimum, outputVoltageRangeMaximum, DAQmx_Val_Volts, NULL ) );
    printf( "Created AO Voltage Chan OK\n" );
    DAQmxErrorCheck ( DAQmxBaseCfgSampClkTiming( outputTaskHandle, clockSource, sampleRate, DAQmx_Val_Rising, DAQmx_Val_ContSamps, samplesPerChannel ) );
    printf( "Set sample rate\n" );
    // trigger output when input starts
    DAQmxErrorCheck ( DAQmxBaseCfgDigEdgeStartTrig( outputTaskHandle, outputTrigger, DAQmx_Val_Rising ) );
    printf( "Set output trigger\n" );
    // ------------------- configuration -----------------------
    // write output signal
    DAQmxErrorCheck ( DAQmxBaseWriteAnalogF64( outputTaskHandle, pointsToWrite, 0, timeout, DAQmx_Val_GroupByScanNumber, outputData, &pointsWritten, NULL ) );
    printf( "Write output signal\n" );
    // set up input buffer
    DAQmxErrorCheck ( DAQmxBaseCfgInputBuffer( inputTaskHandle, 200000 ) ); // use a 100,000 sample DMA buffer
    // initiate acquisition - must start output task first
    DAQmxErrorCheck ( DAQmxBaseStartTask( outputTaskHandle ) );
    DAQmxErrorCheck ( DAQmxBaseStartTask( inputTaskHandle ) );
    // The loop will quit after 10 seconds
    Dr John Clements
    Lead Programmer
    AxoGraph Scientific

    Hi Michael,
    First of all, thanks very much for taking the time to investigate this problem! Much appreciated.
    You asked for "an actual error code you got and any description that is given". The full output from the program that I posted earlier in this thread is appended to the end of this message. In summary, following the call to...
    DAQmxErrorCheck ( DAQmxBaseCfgDigEdgeStartTrig( outputTaskHandle, outputTrigger, DAQmx_Val_Rising ) );
    ... with ...
    char outputTrigger[] = "Dev1/ai/StartTrigger";
    ...the error message is ...
    DAQmxBase Error: Specified route cannot be satisfied, because the hardware does not support it.
    You asked "specifically which M series device you are using"? It is the PCIe 6251 (with BNC 2111 connector block). I'm testing and developing on an Intel Mac Pro (dual boot OS X and Windows XP).
    You asked for "the location you pulled the code from". Here it is...
    http://zone.ni.com/devzone/cda/epd/p/id/879
    ...specifically from the file "Multi-Function-Synch AI-AO_Fn.c".
    I adapted the NI-DAQmx calls to their NI-DAQmx Base equivalents.
    Finally, you asked "Is the trigger necessary, or do you just need to know that the measurements are running on the same clock?". I believe that some kind of sychronized trigger is necessary in my situation (correct me if I'm wrong). Timing is crucial. Say I initiate an analog output stream that delivers a voltage command step 5 ms from the onset. I need to record the response (analog input stream) so that its onset is accurately aligned (synchronized) at 5 ms. A typical recording situation would stimulate and record a short data 'sweep', then wait for the (biological) system to recover, then stimulate and record another short sweep, and repeat. I need all the recorded sweeps to align accurately so that they can be averaged and analyzed conveniently.
    I definitely do not want my customers to rely on an expensive external TTL pulse generator to initiate and synchronize each 'sweep'. That would effectively eliminate the cost advantage of an NI board, as well as adding unnecessary complexity in setup and use. It would be a show-stopper for me.
    It seems perverse, but would it be possible to use a digital output channel connected directly to a digital input chanel to trigger the input and output streams?
    Regards,
    John.
    Full output from test program. Compiled with gcc 4 under OS X...
    [Session started at 2007-05-23 14:17:01 +1000.]
    LoadRuntime: MainBundle
    CFBundle 0x303cc0 (executable, loaded)
    _CompatibleWithLabVIEWVersion: linkedAgainst: 08208002
    _CompatibleWithLabVIEWVersion: result= false, mgErr= 1, theActualVersion= 00000000
    _CompatibleWithLabVIEWVersion: linkedAgainst: deadbeef
    _CompatibleWithLabVIEWVersion: Reseting Linked Against
    _CompatibleWithLabVIEWVersion: linkedAgainst: 08208002
    _CompatibleWithLabVIEWVersion: result= true, mgErr= 0, theActualVersion= 00000000
    _CompatibleWithLabVIEWVersion: linkedAgainst: 08208002
    _CompatibleWithLabVIEWVersion: result= true, mgErr= 0, theActualVersion= 00000000
    com.ni.LabVIEW.dll.nidaqmxbaselv
    CFBundle 0x313760 (framework, loaded)
    {type = 15, string = file://localhost/Library/Frameworks/nidaqmxbaselv.framework/, base = (null)}
    Amethyst:Library:Frameworks:nidaqmxbaselv.framework
    2007-05-23 14:17:02.248 test-ni[4445] CFLog (21): Error loading /Library/Frameworks/LabVIEW 8.2 Runtime.framework/resource/nitaglv.framework/nitaglv: error code 4, error number 0 (no suitable image found. Did find:
    /Library/Frameworks/LabVIEW 8.2 Runtime.framework/resource/nitaglv.framework/nitaglv: mach-o, but wrong architecture)
    CFBundle 0x1751fdc0 (framework, not loaded)
    Created input task
    Created AI Voltage Chan
    Set sample rate
    Created output task
    Created AO Voltage Chan OK
    Set sample rate
    DAQmxBase Error: Specified route cannot be satisfied, because the hardware does not support it.
    test-ni has exited with status 0.
    Dr John Clements
    Lead Programmer
    AxoGraph Scientific

  • Read the file and write the data into string or int

    Ok, here's another rookie question.
    I got to read the line from the file and store it in the memory as string or integer.
    I wrote this code:
    File inputFile = new File("c:/xxx.yyy");
    String outputString = new String();
    FileReader c1 = new FileReader(inputFile);
    StringWriter d1 = new StringWriter(outputString);
    int data;
             while((data = c1.read()) != -1)
                 d1.write(data);
                 System.out.println(data);
             c1.close();
             d1.close();This is how I see it, but it cannot be even compiled because there is no such thing as StringWriter(String) wich I find completely irrational. So tell me, am I retarded or what? No, but seriosly, please help the newbie to understand these streams, am I supposed to use some kind of buffer? Any help appreciated.

    A String is an immutable. You cannot update a String.
    A new String() is basically the same as an empty String ""
    An empty StringWriter can be created with new StringWriter().
    If you use a BufferedReader you can call readLine().
    BufferedReader in = new BufferedReader(new FileReader(filename));
    String line;
    while((line = in.readLine()) != null) {
      System.out.println(line);
    }

Maybe you are looking for

  • Lost emails while transferring to IMAP - What can be done?

    Hello everybody. I'm running 10.8.2 and Mail 6.2. I have a collection of old emails (90K ish) that I recently imported to local Mail storage.   It took me a while to do some sorting but finally got my emails into a few folders. I wanted to upload the

  • JS CS3 script form word styles to indesign styles

    I have search over the forums and can find a script the can change WORD styles (with a diskette sign) to Indesign paragraph styles (without the sign). Can any one help me on this? Thanks.

  • What reigon is a dvd created from new using Toast?

    Hi all I've just made a dvd from some .mov files with no DRM issues using toast titanium. All fine, but I want to send it to someone in New Zealand. I'm told that the dvd will be reigon 0. Is that correct? Thanks in advance

  • IPhone case design suggestion

    The leather cases for $39.95 are good and I have ordered one based on reader recomendations. But I suspect that if they are soft, leaning on them might accidentally activate a function on the iPhone (a freind of mine has already had trouble with this

  • Export Data to Excel Instance

    How to Export the data results to Excel Instance? This feature is available in Toad. Is there a similar functionality with SQL Developer ? Regards Shiva Bodipudi