Serialize and Sequential files.

Hi all :)
This may be a little abstract question:
Is there any difference between any sequential file and a class file which implements the Serializable interface?Or can it be called a sequential file if a .class implements the Serializable interface?
Thanx!

sequential file There were operating systems which had the notion of the "sequential file".
It consisted of records, and they could be acessed one after the other.
There were also index-sequential files too, which could be accessed either sequentially or by keys ("indices").
The OS was aware of these file types.
Modern OS'es (including Unix and Windows) handle the files as byte sequences, leaving the interpretation to the applications.

Similar Messages

  • Differancec between local and sequential files

    hi,
    what is the differancec between local and sequential files

    Sequential files are files with specific format. You upload data into SAP using sequential files.
    ABAP allows you to use sequential files located on the application server or presentation server. You can use these files to buffer data, or as an interface between local programs and the R/3 System
    Sequential files are the files which are stored at application server
    To read them or to put the data into them we use the DATASET concepts for transfering data
    OPEN data set is used to open the file
    READ dataset for reading the file
    TRANSFER dataset for transfering/writing the data
    CLOSE dataset to close the dataset.
    Local file is nothing but ur local system file ..Like C:\, D:\

  • Open all pages in a PDF and save them as sequential files

    I'm trying to make a script that will open all pages in a multi-page PDF and save them as sequential files, but I have had no luck with the simple script in the "Adobe Ilustrator CS5 Reference: JavaScript".
    there example to open a PDF to a designated page does not work for me in Illustrator CS 5.
    // Opens a PDF file with specified options
    var pdfOptions = app.preferences.PDFFileOptions;
    pdfOptions.pDFCropToBox = PDFBoxType.PDFBOUNDINGBOX;
    pdfOptions.pageToOpen = 2;
    // Open a file using these preferences
    var fileRef = filePath;
    if (fileRef != null) {
    var docRef = open(fileRef, DocumentColorSpace.RGB);
    if anyone can get me on the right path it would be appreciated.

    Hi DuanesSFK,
    you can open a PDF page on this way (tested with CS3 and 5):
    //OpenPDFpage2.jsx
    var uILevel = userInteractionLevel; 
    app.userInteractionLevel = UserInteractionLevel.DONTDISPLAYALERTS;
    // Opens a PDF file with specified options
    var pdfOptions = app.preferences.PDFFileOptions;
    pdfOptions.pDFCropToBox = PDFBoxType.PDFBOUNDINGBOX;
    pdfOptions.pageToOpen = 2;
    // Open a file using these preferences
    //var fileRef = filePath;
    var fileRef = File("~/Desktop/YourOwn.pdf");
    if (fileRef != null) {
    var docRef = open(fileRef, DocumentColorSpace.RGB);
    app.userInteractionLevel = uILevel;
    To open all pages you have to loop through all pages etc.
    Easier for you:
    Search here in forum for the "open pdf" script  (I don't know the exactly name, it was written by CarlosCanto, ) This should always work for you.
    Have fun

  • Very high log file sequential read and control file sequential read waits?

    I have a 10.2.0.4 database and have 5 streams capture processes running to replicate data to another database. However I am seeing very high
    log file sequential read and control file sequential read by the capture procesess. This is causing slowness in the database as the databass is wasting so much time on these wait events. From AWR report
    Elapsed: 20.12 (mins)
    DB Time: 67.04 (mins)
    and From top 5 wait events
    Event Waits Time(s) Avg Wait(ms) % Total Call Time Wait Class
    CPU time 1,712 42.6
    log file sequential read 99,909 683 7 17.0 System I/O
    log file sync 49,702 426 9 10.6 Commit
    control file sequential read262,625 384 1 9.6 System I/O
    db file sequential read 41,528 378 9 9.4 User I/O
    Oracle support hasn't been of much help, other than wasting my 10 days and telling me to try this and try that.
    Do you have streams running in your environment, are you experiencing this wait. Have you done anything to resolve these waits..
    Thanks

    Welcome to the forums.
    There is insufficient information in what you have posted to know that your analysis of the situation is correct or anything about your Streams environment.
    We don't know what you are replicating. Not size, not volume, not type of capture, not rules, etc.
    We don't know the distance over which it is being replicated ... 10 ft. or 10 light years.
    We don't have any AWR or ASH data to look at.
    etc. etc. etc. If this is what you provided Oracle Support it is no wonder they were unable to help you.
    To diagnose this problem, if one exists, requires someone on-site or with a very substantial body of data which you have not provided. The first step is to fill in the answers to all of the obvious first level questions. Then we will likely come back with a second level of questioning.
    But when you do ... do not post here. Your questions are not "Database General" they are specific to Streams and there is a Streams forum specifically for them.
    Thank you.

  • Read data from a sequential file to fill an arrray

    I am trying to read data from a sequential file to fill an array and then use the array for calculations. I believe my program is reading the file; however when I try to use the data for calculations the results show up as boxes rather than numbers. Can someone please take a look and give me some insight into where I am going wrong. The sequential file has the following data:
    5.0
    5.35
    5.5
    5.75
    Here is my code from that portion of the program. I can send the entire program if necessary.
    private void calcResults(TextArea loanResults, double amount, double interest, int term ) throws IOException {
         DecimalFormat df = new DecimalFormat("$###,###.00");
         NumberFormat formats = new DecimalFormat("#0.00");
         StringBuffer buffer=new StringBuffer();
         loanResults.append("Month No.\tMonthly Payment\t\tInterest\t\tBalance\n");
         double monthlyPay = amount*Math.pow(1+interest,term)*interest/(Math.pow(1+interest,term)-1);
         monthlyPayment.setText("" + (formats.format(monthlyPay)));
         double principal= amount;
          * Loop through each month of a given loan
        int month;
         for (int i=0; i<term; i++)
    try {
         int j = 0;
         BufferedReader in = new BufferedReader(new FileReader("rates.txt"));
         String temp = "";
         while((temp = in.readLine()) != null) {     
            j++;
            strRate[j] = temp;
            RateValue1 = Double.parseDouble(loanAmountTxFld.getText());
            in.close();
             catch (FileNotFoundException e) {
                 System.out.println("Can't find file rate.txt!");
                 return;
          month= i+1;
          double rate =principal*interest;
          double balance=principal+rate-monthlyPay;
          loanResults.append((month) + "\t\t" + (formats.format(principal) + "\t\t"  + (formats.format(rate) + "\t\t"
                            + (formats.format(balance)     + "\n"))));
          principal=balance;
          GraphArea.append(formats.format(balance)     + "\n");
    *Method for determining which loan option was chosen
    private void calc() throws IOException{
      String interestTerms = (String) cOption.getSelectedItem();
      if (interestTerms.equalsIgnoreCase("5 yrs at 5.00"))
         calcResults(loanResults, Double.parseDouble(loanAmountTxFld.getText()), 0.05/12, 60);
      else if (interestTerms.equalsIgnoreCase("7 yrs at 5.35"))
        calcResults(loanResults, Double.parseDouble(loanAmountTxFld.getText()), RateValue1/12, 84);
      else if (interestTerms.equalsIgnoreCase("15 yrs at 5.50"))
         calcResults(loanResults, Double.parseDouble(loanAmountTxFld.getText()), 0.0550/12, 180);
      else if (interestTerms.equalsIgnoreCase("30 yrs at 5.75"))
         calcResults(loanResults, Double.parseDouble(loanAmountTxFld.getText()), 0.0575/12, 360);
      else if (interestTerms.equalsIgnoreCase("       "))
        calcResults(loanResults, Double.parseDouble(loanAmountTxFld.getText()), Double.parseDouble(loanInterestTxFld.getText())/100/12,Integer.parseInt(loanTermTxFld.getText())*12);
    }

    ok, I fixed my program per your suggestion and I still cannot get it to work. I get the same result where the ouput is just a bunch of boxes. I also tried to print to see what I am getting and all I see is the values of the array [5.0,5.35,5.5,5.75] but I cannot seem to pass that to the calculations. I wish I could figure this out. Does anybody have any suggestions as to what I am doing wrong. This is the portion of my code that I am having issues with:
    private void calcResults(TextArea loanResults, double amount,double interest, int term ) throws IOException {
         DecimalFormat df = new DecimalFormat("$###,###.00");
         NumberFormat formats = new DecimalFormat("#0.00");
         StringBuffer buffer=new StringBuffer();
         loanResults.append("Month No.\tMonthly Payment\t\tInterest\t\tBalance\n");
         double monthlyPay = amount*Math.pow(1+interest,term)*interest/(Math.pow(1+interest,term)-1);
         monthlyPayment.setText("" + (formats.format(monthlyPay)));
         double principal= amount;
          * Loop through each month of a given loan
        int month;
         for (int i=0; i<term; i++)
          month= i+1;
          double rate =principal*interest;
          double balance=principal+rate-monthlyPay;
          loanResults.append((month) + "\t\t" + (formats.format(principal) + "\t\t"  + (formats.format(rate) + "\t\t"
                            + (formats.format(balance)     + "\n"))));
          principal=balance;
          GraphArea.append(formats.format(balance)     + "\n");
    private void readFile() throws IOException
    try {
         int j = 0;
         BufferedReader in = new BufferedReader(new FileReader("rates.txt"));
         String temp = "";
         while((temp = in.readLine()) != null) {     
            strRate[j] = temp;
            j++;
    RateValue1 = Double.valueOf(((String)(strRate[1]))
    ).doubleValue();
    //        RateValue1 = Double.parseDouble(loanAmountTxFld.getText());
            in.close();
             catch (FileNotFoundException e) {
                 System.out.println("Can't find file rates.txt!");
                 return;
    *Method for determining which loan option was chosen
    private void calc() throws IOException{
      String interestTerms = (String) cOption.getSelectedItem();
      if (interestTerms.equalsIgnoreCase("5 yrs at 5.00"))
         calcResults(loanResults, Double.parseDouble(loanAmountTxFld.getText()), 0.05/12, 60);
      else if (interestTerms.equalsIgnoreCase("7 yrs at 5.35"))
        calcResults(loanResults, Double.parseDouble(loanAmountTxFld.getText()), RateValue1/12, 84);
      else if (interestTerms.equalsIgnoreCase("15 yrs at 5.50"))
         calcResults(loanResults, Double.parseDouble(loanAmountTxFld.getText()), 0.0550/12, 180);
      else if (interestTerms.equalsIgnoreCase("30 yrs at 5.75"))
         calcResults(loanResults, Double.parseDouble(loanAmountTxFld.getText()), 0.0575/12, 360);
      else if (interestTerms.equalsIgnoreCase("       "))
        calcResults(loanResults, Double.parseDouble(loanAmountTxFld.getText()), Double.parseDouble(loanInterestTxFld.getText())/100/12,Integer.parseInt(loanTermTxFld.getText())*12);
      }

  • Help required for slicing and merging file

    Hi,
    I am working on p2p downloading application based on Gnutella. Right now i am trying to implement downlaoding one file from many hosts at a time(like limewire does). i am able to connect to multiple host. but dont know how to downlaod and merge file for his purpose.
    i know how to download one file from single host. but dont know how to handle file coming from multiple hosts.
    if anyone having knowledge, please tell as soon as possble.

    Hi.
    thanks for the reply. Actually i had something same
    in my mind. but unfortunately i am quite week in
    serializtion and file stuff. If you could provide me
    with code snipet for this, or can just simulate in a
    fewer lines of code, it would be a great help for
    me...i have too much short time...
    i think i really have to study this file
    stuff..........You shouldn't use serialization. Just invent a protocol, and "normal" binary data.
    Kaj

  • Location of stubs, skeletons, remote interface ,client and server files

    I have a question to the subject mentioned above :
    In a training example for RMI I red the following :
    .... stubs, remote interface and client files into
    client location
    ... skeletons, remote interfaces, server files and stubs
    into server location ....
    I understand that with the exception that the stubs must also be located at server side. My understanding is
    that it is enough if those reside only at client side. Is that
    right or wrong ?
    Thanks for your comment.
    Regards,
    Oskar

    The stubs must be available (in the classpath) on the server side. This is because when a remote server object is exported, the RMI runtime creates an instance of the stub (instead of the server) and sends it whenever a reference to the server is encountered. In order to create (and then serialize) the stub, its class file must be available on the server.

  • How to serialize a pdf file?

    A pdf file is made up of primitive and non-primitive data i need to serialize a pdf file and send it to another computer via rmi. I cannot find any examples on how to go about it. Can anybody help?

    As bytes.

  • User I/O and db file parallel read is high in AWR report

    Hi,
    We have one performance issue during a job execution.
    From the awr report we have identified one query with a table having millions of records causing problems and then we had also fine tuned that query by changing it's code and by using the optmizer hints. It is being executed in plsql batches. After fine tuning, On the first batch execution(first 5000 records) the query is taking only 5 mins, but on the consecutive batches it is consuming more time( more than 30 mins).
    From the awr report I got the statistics as
    Release : 11.2.0.2.0
    Instance Efficiency Percentages (Target 100%)
    Buffer Nowait %: 100.00 Redo NoWait %: 100.00
    Buffer Hit %: 85.44 In-memory Sort %: 99.98
    Library Hit %: 99.76 Soft Parse %: 99.15
    Execute to Parse %: 88.91 Latch Hit %: 100.00
    Parse CPU to Parse Elapsd %: 87.32 % Non-Parse CPU: 98.65
    The buffer hit % is good. On each batch execution it is taking different set of records.
    Top 5 Timed Foreground Events
    Event Waits Time(s) Avg wait (ms) % DB time Wait Class
    db file parallel read 120,485 42,540 353 89.60 User I/O
    DB CPU 3,794 7.99
    db file sequential read 145,074 606 4 1.28 User I/O
    db file scattered read 70,030 556 8 1.17 User I/O
    direct path write temp 12,423 21 2 0.04 User I/O
    So the I/O is our main concern since that query contains one table with millions of records.
    Host CPU (CPUs: 24 Cores: 24 Sockets: 4)
    Load Average Begin Load Average End %User %System %WIO %Idle
    1.40 1.45 0.6 0.3 3.7 99.0
    Load is also normal.
    From the Time model statistics , sql execute elapsed time is 98.27% of db time and only 7.99% is that of DB CPU.
    Memory Statistics
    Begin End
    Host Mem (MB): 64,318.0 64,318.0
    SGA use (MB): 30,720.0 30,720.0
    PGA use (MB): 488.2 497.1
    % Host Mem used for SGA+PGA: 48.52 48.54
    Both the size of sga_max_size and sga_target are 32,212,254,720(32gb) bytes and that of
    pga_aggregate target is 629,145,600(600mb)
    from this it is evident that the memory is still available(so increase in memory size is not an option).
    The sql statistics for that query shows like that
    Elapsed Time (s) Executions Elapsed Time per Exec (s) %Total %CPU %IO SQL Id SQL Text
    44,950.03 55 817.27 94.67 6.99 94.72 79dgmrxh4kv74 SELECT /*+ index(cdr_data cdr_...
    I can't understand whether the problem is in the database side or with the query?
    If the problem is with the query, then how it has been executed in 5 mins for the first batch ?
    (all the batches are having 5000 records each).
    And how can we reduce the db file parallel read ?
    Your valuable advice will be greatly appreciated.
    Thanks in advance
    Manoj Kumar N

    "db file parallel read" is likely to be associated with something like index prefetching.
    See:
    http://www.freelists.org/post/oracle-l/RE-Calculating-LIOs,11
    http://aprakash.wordpress.com/2012/05/29/index-range-scan-and-db-file-scattered-read-as-session-wait-event/
    http://jonathanlewis.wordpress.com/2006/12/15/index-operations/
    Tune the SQL.
    Review the execution plan.
    Check whether the statistics are accurate.
    Review whether the index hint (and others that we can't see) is appropriate.

  • Multiplexing redo logs and control files to a separate diskgroup

    General question this one...
    I've been using ASM for a few years now and have always installed a new system with 3 diskgroups
    +DATA - for datafiles, control files, redo logs
    +FRA - for achive logs, flash recovery. RMAN backup
    Those I guess are the standards, but I've always created an extra (very small) diskgroup, called +ONLINE where I keep multiplexed copies of the redo logs and control files.
    My reasoning behind this is that if there are any issues with the +DATA diskgroup, the redo logs and control files can still be accessed.
    In the olden days (all those 5 years ago!), on local storage, this was important, but is it still important now? With all the striping and mirroring going on (both at ASM and RAID level), am I just being overtly paranoid? Does this additional +ONLINE diskgroup actually hamper performance? (with dual write overheads that are not necessary)
    Thoughts?

    Some of the decision will probably depend on your specific environment's data activity, volume, and throughput.
    Something to remember is that redo logs are sequential write, which benefit from a lower RAID overhead (RAID-10, 2 writes per IOP vs RAID-5, 4 writes per IOP). RAID-10 is often not cost-effective for the data portion of a database. If your database is OLTP with a high volume of random reads/writes, you're potentially hurting redo throughput by creating contention on the disks sharing data and redo. Again, that depends entirely on what you're seeing in terms of wait events. A low volume database would probably not experience any noticeable degraded performance.
    In my environment, I have RAID-5 and RAID-10 available, and since the RAID-10 requirement from a capacity perspective for redo is very low, it makes sense to create 2 diskgroups for online redo, separate from DATA, and separate from each other. This way, we don't need to be concerned with DATA transactions impacting REDO performance, and vice versa, and we still maintain redo redundancy.
    In my opinion, you can't be too paranoid. :)
    Good luck!
    K

  • Getting dump when passing to sequential file

    hi to all,
    pla help me in this issue
    in final internal table iam having to currency field
    i want to pass to sequential file but it is going to dump
    it is giving this error
    Only character-type data objects are supported at the argument
    position "f" for the statement
      "TRANSFER f TO ...".
    In this case, the operand "f" has the non-character-type "ZPAYREF". The
    current program is flagged as a Unicode program. In the Unicode context
    type X fields are seen as non-character-type, as are structures that
    contain non-character-type components.
    plz help how to passs amount final currency field like character so that i can move to sequentail file.
    Thanks and Regards
    Kranthi

    Hi,
    Data : l_qty(16) type c,
           l_value(13) type c,
           l_dec(3)   type c.
    loop at itab.
    split itab-qty at `.` into l_value l_dec.
    concatenate l_value `.` l_dec into l_qty.
    Move l_qty to l_file.
    Transfer l_file.
    endloop.
    close dataset.
    Best regards,
    Prashant

  • Singleton serialization and weblogic - help!

    hi everyone,
    I'm trying to serialize a singleton cache object we're using for our (intranet)
    website running on weblogic 6.1. I've made sure the singleton class
    implements serializable, and i've also put the following method in the
    singleton class (ChartCache.java):
    private Object readResolve()
    return theInstance; //which is of type ChartCache
    By the way I'm using jdk 1.3. I've trawled the web and can't find anything to help me yet...
    Basically I've written a jsp to get the current instance of the cache
    and serialize it to a file. When I invoke another jsp i want it to
    reinflate the object and place it back into the weblogic jvm so it can
    once again be accessed as a singleton by all classes in the webapp. At
    the moment its just not working - the cache object will eventually
    take over half an hour to create (huge database processing going on)
    so I want the option, if weblogic falls over, to reinstantiate the
    cache without having to rebuild it from scratch. The cache will only
    be refreshed once a day.
    I'll c&p my jsp code below...
    ======================= restoreCache.jsp
    <%@ page import = "java.util.*" %>
    <%@ page import = "java.io.*" %>
    <%@ page import = "com.drkw.agencylending.website.chartcache.*" %>
    <%@ page import = "com.drkw.agencylending.log.Logger" %>
    <%
    ChartCache theCache = null;
    try
    String filename = "chartCache.dat";
    FileInputStream fis = new FileInputStream(filename);
    ObjectInputStream in = new ObjectInputStream(fis);
    theCache = (ChartCache)in.readObject();
    in.close();
    catch(Exception e)
    Logger.log(e);
    %>
    Resurrected chart cache from file.
    ======================= saveCache.jsp
    <%@ page import = "java.util.*" %>
    <%@ page import = "java.io.*" %>
    <%@ page import = "com.drkw.agencylending.website.chartcache.*" %>
    <%@ page import = "com.drkw.agencylending.log.Logger" %>
    <%
    ChartCache theCache = ChartCache.getInstance();
    String fileName = "chartCache.dat";
    try
    FileOutputStream fos = new FileOutputStream(fileName);
    ObjectOutputStream oos = new ObjectOutputStream(fos);
    oos.writeObject(theCache);
    oos.flush();
    oos.close();
    catch(Exception e)
    Logger.log(e);
    %>
    Wrote cache object to file.
    ============================
    At the moment, a file is getting written but it seems suspiciously
    small (73 bytes!) for what is a very large object. When I try to load
    it up again, I get no errors but when I call ChartCache.getInstance()
    (my singleton) it recreates the cache rather than using the one I've
    reinflated. BTW I don't need to worry about server clustering/JNDI to
    ensure I have a truly singleton instance - there will only ever be one
    JVM to worry about.
    Any help greatly appreciated!

    solved the problem using another technique. Rather
    than serialize the entire singleton object, I just
    serialized the hashmap (the only bit I really care
    about). Then in the private constructor, the class
    checks a database flag to see whether the singleton
    was created today - if not, I refresh the cache,
    otherwise load in the serialized hashmap and set that
    as my class member. No complicated jndi/ejb registry
    needed to get around it..!singleton serialization is not recommended in ejb enviroment, it will cause problems. in what you are doing, it is far better to persist the data values in a database.
    just for the sake of writing a serializable sington: in addition to what you do in a regular object, you would first check to see if this obj has been serialized, if it is, deserialize it and return this one. you would have to write your own method to do serialization, making sure you have only one place to serialize it, and let the object know where the file is.

  • Reg sequential file

    hi
    i want to know about sequential file in data transfer.wat is the concept behind this.
    thanks in advance

    I assume you are clear about the concepts of Application Server and the Presentation server.
    some times the data that has to be transfered in SAP comes in the form of Files. So these files have to be read, processed (if required) and then have to be passed either to transactions or some Function Modules.
    These files can be on App Server or on Presentation server.
    To read file from App server you do Open dataset, close dataset and read data set.
    To read files form Pres Server... there are FMs available GUI_UPLOAD and GUI_DOWNLOAD.
    This is the general.. If u want some thing in specific.. Please feel free to ask...

  • How to use jsp and xhtml files in an application

    This is web.xml file,
    <web-app>
    <context-param>
         <param-name>javax.faces.DEFAULT_SUFFIX</param-name>
         <param-value>.xhtml</param-value>
    </context-param>
    <context-param>
    <param-name>facelets.DEVELOPMENT</param-name>
    <param-value>true</param-value>
    </context-param>
    <servlet>
         <servlet-name>Faces Servlet</servlet-name>
         <servlet-class>javax.faces.webapp.FacesServlet</servlet-class>
         <load-on-startup>1</load-on-startup>
    </servlet>
    <servlet-mapping>
         <servlet-name>Faces Servlet</servlet-name>
         <url-pattern>*.jsf</url-pattern>
    </servlet-mapping>
    <listener>
    <listener-class>org.apache.myfaces.webapp.StartupServletContextListener</listener-class>
    </listener>
    <context-param>
    <param-name>javax.faces.CONFIG_FILES</param-name>
    <param-value>/WEB-INF/faces-config.xml</param-value>
    </context-param>
    </web-app>
    If i configure like this, i am not able to work with jsp file.
    So, how to configure web.xml file to work with both jsp and xhtml file.
    Thanks,
    Vinutha.

    Hi Sam,
    To Use properties file you need to keep the properties file in a location and in the code you need to mention the path from where it should read in the runtime.
    Fuego.Io.PropertiesFile propfile;
    propfile.load(fileName : "<Path: C:/sample.properties>"); //Path: Place the properties file in the respective directory/drive and mention the path
    String val = propfile.get(key : "One"); //One - is the key against which the value should be mentioned in the properties file like <One = 1>. It should return 1
    logMessage("Value: " + val);
    To use the Enum in your project first create a Module in your catalogue component and then right click on module create New Enumeration say ProcessStatus
    Uncheck the Is Sequential if you want to keep key/value as pair. Click on add +* and mention the Name and Value say ABORTED as Name and Value as Aborted. To use it in your project conditional path write the condion as somevariable == String(ProcessStatus.ABORTED)
    Hope this will help you.
    Bibhu

  • Rman tivoli failure:  create sequential file, Authentication Failure

    New db instance, not writing to tape.
    DB is 10.2.0.4
    run {
    allocate channel sbt_backup1 type 'SBT_TAPE' format '%U' parms 'ENV=(TDPO_OPTFILE=/usr/tivoli/tsm/client/ora
    cle/bin64/tdpo_mydb.opt)';
    backup tablespace users;
    release channel sbt_backup1;
    RMAN-00571: ===========================================================
    RMAN-00569: =============== ERROR MESSAGE STACK FOLLOWS ===============
    RMAN-00571: ===========================================================
    RMAN-03009: failure of backup command on sbt_backup1 channel at 10/02/2009 11:17:42
    ORA-19506: failed to create sequential file, name="0tkqpi6l_1_1", parms=""
    ORA-27028: skgfqcre: sbtbackup returned error
    ORA-19511: Error received from media manager layer, error text:
    ANS1025E (RC137) Session rejected: Authentication failure
    Settings
    /usr/tivoli/tsm/client/oracle/bin64
    dsm_mydb.opt
    SErver_name MYDBPRD
    tdpo_mydb.opt
    DSMI_ORC_CONFIG /usr/tivoli/tsm/client/oracle/bin64/dsm_mydb.opt
    TDPO_NODE MYDBPRD
    /usr/tivoli/tsm/client/api/bin64
    dsm.sys
    NODENAME MYDBPRD
    ERRORLOGNAME /usr/local/tsm/logs/IFSDBS_MYDBPRD_error.log
    SCHEDLOGNAME /usr/local/tsm/logs/IFSDBS_MYDBPRD_sched.log
    Note: Several db/clients are working on this server. This is a new setup and not working.

    This is not an Oracle/RMAN problem,contact your TSM administrator:
    http://www-01.ibm.com/support/docview.wss?uid=swg21216057
    Werner

Maybe you are looking for

  • Adobe wont allow me to renew my license

    Hey, so, my company CC got cancelled, and ever since it wont allow me to use a difference card. Customer support just send me round in circles always promising to call back, but never do. Has anyone ever had a similar problem / solution? I deseperate

  • I want recover in ABAP program a "Long Text"

    Hi,  In FB01 transaction when you go to "Enter Fixed Asset posting: Correct G/L account item", there are a button with a label: "Long Text". When you push this button appear a little window with name "Long Text for Document Line Item" and you can wri

  • Text engine won't initalize in CS5

    I had no problems with CS5 until today.  I downloaded a trial version of In Design.  Now, the text engine in CS5 won't initialize.  What can I do?  I tried to call the phone support and he kept trying to sell me CS6.  I got so frustrated I hung up! 

  • TS4062 ipod classic issues when trying to sync with itunes

    i have an ipod classic which has always worked great when getting music from itunes, but i bought a new computer and now i can't get it to sync. i also have an ipod touch which works fine. i need help figuring out how to sync with my new computer and

  • Swf file transparent in Director MX

    Hello: Do you know how to get an animation in Flash with background transparent? If I put ink:background transparent does not work. Thank you