Query Result Replacement between different objects

Hi All
I have a req. where i need to pass result of one query and restrict the second query with this. Usually its easy if they both are same object like say Sales Document(0DOC_NUMB). But in my case i need to pass 0DOC_NUMB from first query to 0REF_DOC in second query. Has anyone done this? Any help will be appreciated.
Thanks

Hi ,
You can try creating a replacement path variable on 0DOC_NUMB getting filled by the first query.
Then you can pass these values to a variable of type customer exit on 0REF_DOC in second query in the user exit for variables.
I guess however both the variables need to be used in the second query to get the results.
rgds
Deepak

Similar Messages

  • Cache an LDAP query result in a Map Object

    Is there a way to perform a single LDAP query and store it in some type of an indexed list Object in memory. Specifically I need to populate both LDAP manager and managerFullName for an LDAP user object based on an employeenumber query.
    I don't want to query LDAP for every user object. I would like to submit one search such as (objectclass=inetorgperson) and store the result in an indexed list in memolry using employeenumber as the key. This way I only need to query the indexed list object for each user entry.
    Is this possible?

    No this is not possible.
    The only way to do this is to use a java class you write yourself. But and a major but: if you do not stay in the same place in IDM (form or workflow) you will lose the content because the object will be garbage collected when you change.
    The other thing is: how much will you gain? The ldap server can probably return the result far quicker then you can iterate through the list to find the entry.
    WilfredS

  • Relation between different objects

    Dear Experts,
    What is the interlink between Material or acivity to cost-elements, cost components , GL accounts and how does each objects influence costing.

    All these transactions when you are posted in Controlling requires a Cost element (Primary or Secondary), GL Account in Finance. Futher to display the costs in Cost component Structure, each cost component is assinged to cost elements.

  • Iterate query result - put specific rows in a collection and do output

    JDeveloper 11.1.1.3.0
    Oracle Lite: 10.2.0.1.0
    Browser: IE 7
    Hi,
    please help me in working on following requirement:
    1. Iterate the View Object Query Result
    2. Do some calculation on the Rows, lets say:
    Concatination of column1 and column2 and place it in a non database Attribute "column3"
    3. How to handle/place the non database attribute?
    4. Fetch specific rows from the query result; there should be a comparison, like checking master and detail hierachy struktur:
    Take Column1_ID (Master) check in columns2_ID (Detail), when found detail take this Row and place it in a buffer (Maybe ArrayList, or something like that)
    5. Show the Chace(ArrayList) to user, in a form of View Object, with standard functions like sorting and filtering
    Till now my process of solving the above requiremnts are:
    To Point 1:
    Created a service method in Application Implementation Class
       public void iterateEmp(){
         EmployeesVOImpl vo = getEmployeesVO1();
         while (vo.hasNext()){
          EmployeesVORowImpl row = (EmployeesVORowImpl)vo.next();
          System.out.println(row.getEmployeeId() + " | " + row.getFirstName() + " | " + row.getLastName() + " | " +  row.getEmail());   
          //row.setCalcCol(row.getFirstName() + "," + row.getLastName());
         }To point 2:
    Added an attribute to the View Object, with updateable and Queryable options. Place it on the JSPX Page.
    With setter method, filled with values, for example (non Database Attributes Name is CalcCol):
    Name:
    //see above code
    row.setCalcCol(row.getFirstName() + "," + row.getLastName());When doing so, on initial Call the values will be representated, but when sorting o filtering will be done, the values then will be erased.
    3. Added an Attribute in View Object "Attributes Panel".
    Here the XML Code of the View Object:
    <?xml version="1.0" encoding="windows-1252" ?>
    <!DOCTYPE ViewObject SYSTEM "jbo_03_01.dtd">
    <!---->
    <ViewObject
      xmlns="http://xmlns.oracle.com/bc4j"
      Name="EmployeesVO"
      Version="11.1.1.56.60"
      BindingStyle="OracleName"
      CustomQuery="true"
      ComponentClass="model.EmployeesVOImpl"
      PageIterMode="Full"
      UseGlueCode="false"
      RowClass="model.EmployeesVORowImpl">
      <DesignTime>
        <Attr Name="_codeGenFlag2" Value="Access|Coll|VarAccess"/>
        <Attr Name="_isExpertMode" Value="true"/>
        <Attr Name="_isCodegen" Value="true"/>
      </DesignTime>
      <SQLQuery>
        <![CDATA[select employee_id, first_name, last_name, email from employees]]>
      </SQLQuery>
      <ViewAttribute
        Name="EmployeeId"
        IsUpdateable="false"
        IsPersistent="false"
        IsNotNull="true"
        PrecisionRule="true"
        Precision="6"
        Scale="0"
        Type="oracle.jbo.domain.Number"
        ColumnType="NUMBER"
        AliasName="EMPLOYEE_ID"
        Expression="EMPLOYEE_ID"
        SQLType="NUMERIC">
        <DesignTime>
          <Attr Name="_DisplaySize" Value="22"/>
        </DesignTime>
      </ViewAttribute>
      <ViewAttribute
        Name="FirstName"
        IsUpdateable="false"
        IsPersistent="false"
        PrecisionRule="true"
        Precision="20"
        Type="java.lang.String"
        ColumnType="VARCHAR2"
        AliasName="FIRST_NAME"
        Expression="FIRST_NAME"
        SQLType="VARCHAR">
        <DesignTime>
          <Attr Name="_DisplaySize" Value="20"/>
        </DesignTime>
      </ViewAttribute>
      <ViewAttribute
        Name="LastName"
        IsUpdateable="false"
        IsPersistent="false"
        IsNotNull="true"
        PrecisionRule="true"
        Precision="25"
        Type="java.lang.String"
        ColumnType="VARCHAR2"
        AliasName="LAST_NAME"
        Expression="LAST_NAME"
        SQLType="VARCHAR">
        <DesignTime>
          <Attr Name="_DisplaySize" Value="25"/>
        </DesignTime>
      </ViewAttribute>
      <ViewAttribute
        Name="Email"
        IsUpdateable="false"
        IsPersistent="false"
        IsNotNull="true"
        PrecisionRule="true"
        Precision="25"
        Type="java.lang.String"
        ColumnType="VARCHAR2"
        AliasName="EMAIL"
        Expression="EMAIL"
        SQLType="VARCHAR">
        <DesignTime>
          <Attr Name="_DisplaySize" Value="25"/>
        </DesignTime>
      </ViewAttribute>
      <ViewAttribute
        Name="CalcCol"
        IsSelected="false"
        IsPersistent="false"
        PrecisionRule="true"
        Type="java.lang.String"
        ColumnType="VARCHAR2"
        AliasName="VIEW_ATTR"
        SQLType="VARCHAR"/>
      <AttrArray Name="KeyAttributes">
        <Item Value="EmployeeId"/>
      </AttrArray>
    </ViewObject>To Point 4:
    There should be a specific sort order, which i have to code in ADF.
    Just to give an overview:
    Colum1 and Column2 (nothing to do with the emp table) have relationship. Several hierachies are existing in the Query Result.
    For identifying the start point of any of these hierachies, i have to iterate through the Query Result and fetch these start point
    and buffer it in an array or something like that.So in the end i have to get the same query Result but with different Sort Order.
    How to buffer theses Rows (Array?) and show in a View Object?
    Can we use the same view Object for solving that, or should we use another View Object which catches the new Result Set?
    To Point 5:
    The Result should be, as mentioned above, a View Object including new Result Set, which can provide standard behavior (sort, filter)
    I hope the requirement is defined clearly enough, so maybe you can make some suggestions on that.
    Thank you for replies.
    Edited by: user9198377 on 13.07.2011 02:52

    Hi.. You can do this without DB View. Have CalcCol transient attribute in VO. Instead of iterating VO rows and setting CalcCol attr, generate View Row Impl class. Inside getCalcCol(), write your logic. This will be called during each and every request so that transient attr data won't become null during sorting and filtering. If your VO is entity driven, do the same in Entity impl class
    Raghu
    Edited by: Raguraman on Sep 5, 2011 9:48 AM

  • Export query results to flat file with dynamic filename

    Hi
    Can anybody can point me how to dynamic export query serults set to for example txt file using process flows in OWB.
    Let say I have simple select query
    select * from table1 where daterange >= sysdate -1 and daterange < sysdate
    so query results will be different every day because daterange will be different. Also I would like to name txt file dynamicly as well
    eg. results_20090601.txt, results_20090602.txt, results_20090603.txt
    I cant see any activity in process editor to enter custom sql statment, like it is in MSSQL 2000 or 2005
    thanks in advance

    You can call existing procedures from a process flow the procedure can create the filename with whatever name you desire. OWB maps with file as target can also create a file with a dynamic name defined by an expression (see here ).
    Cheers
    David

  • Howto handle persistence API query result list

    Hello.
    I need some advice on how to handle a persistence API query result List since the Object variable returned by the "list.get(i)" method cannot be cast to an EntityClass declared by the EntityClassManager.
    The piece of code below might help you understand my exact problem.
    EntityManagerFactory emf = Persistence.createEntityManagerFactory("SporTakipPU");
    EntityManager em = (EntityManager) emf.createEntityManager();
    Query q1 = em.createQuery( "SELECT d.id, d.date, c.name, c.sirname FROM Log d, Kisi c WHERE d.id = :id AND c.id = :id" );
    q1.setParameter( "id", argument );
    List logList = q1.getResultList();
    Log log = (Log) logList.get(1);
    In this code i cannot cast Object to Log in the last line. Apparently this has something to do with the query.
    Trying to solve this problem i declared two private string variables name and sirname and their set and get methods but this didn't help.
    Thanks...

    Entity Class : ReportView
    ===================
    @Entity
    @Table(name = "dummy")
    public class ReportView implements Serializable {
    @Id
    @Column(name = "id", nullable = false)
    private Integer id;
    // getter , setter
    SessionBean : ReportSessionBean
    ==========================
    public List<ReportView> getReports() {
    Query q = em.createNativeQuery("SELECT ID FROM REPORT", ReportView.class);
    List<ReportView> r=(List<ReportView>)q.getResultList();
    return r;
    JSF / Back Bean : ReportBackBean
    ==========================
    public class ReportBackBean {
    @EJB
    private ReportSessionRemote reportSessionBean;
    /** Creates a new instance of ReportBackBean */
    public ReportBackBean() {
    public List<ReportView> getReports() {
    return reportSessionBean.getReports();
    JSF
    ===
    <f:view>
    <h:dataTable value="#{ReportBackBean.reports}" var="report" border="1">
    <h:column>
    <h:outputText value="#{report.id}"/>
    </h:column>
    </h:dataTable>
    </f:view>
    Regards,
    Telman
    ..

  • Stepping through a query result set, replacing one string with another.

    I want to write a function that replaces the occurance of a string with another different string.  I need it to be a CF fuction that is callable from another CF function.  I want to "hand" this function an SQL statement (a string) like this:   (Please note, don't bother commenting that "there are eaiser ways to write this SQL..., I've made this simple example to get to the point where I need help.  I have to use a "sub_optimal" SQL syntax just to demonstrate the situation)
    Here is the string I want to pass to the function:
    SELECT
      [VERYLONGTABLENAME].FIRST_NAME,
      [VERYLONGTABLENAME].LAST_NAME,
      [VERYLONGTABLENAME].ADDRESSS
    FROM
      LONGTABLENAME [VERYLONGTABLENAME]
    Here is the contents of the ABRV table:
    TBL_NM,  ABRV    <!--- Header row--->
    VERYLONGTABLENAME, VLTN
    SOMEWHATLONGTALBENAME, SLTN
    MYTABLENAME, MTN
    ATABLENAME, ATN
    The function will return the original string, but with the abreviations in place of the long table names, example:
    SELECT
      VLTN.FIRST_NAME,
      VLTN.LAST_NAME,
      VLTN.ADDRESSS
    FROM
      LONGTABLENAME VLTN
    Notice that only the table names surrounded by brackets and that match a value in the ABRV table have been replaced.  The LONGTABLENAME immediately following the FROM is left as is.
    Now, here is my dum amatuer attempt at writing said function:  Please look at the comment lines for where I need help.
          <cffunction name="AbrvTblNms" output="false" access="remote" returntype="string" >
            <cfargument name="txt" type="string" required="true" />
            <cfset var qAbrvs="">  <!--- variable to hold the query results --->
            <cfset var output_str="#txt#">  <!--- I'm creating a local variable so I can manipulate the data handed in by the TXT parameter.  Is this necessary or can I just use the txt parameter? --->
            <cfquery name="qAbrvs" datasource="cfBAA_odbc" result="rsltAbrvs">
                SELECT TBL_NM, ABRV FROM BAA_TBL_ABRV ORDER BY 1
            </cfquery>
         <!--- I'm assuming that at this point the query has run and there are records in the result set --->
        <cfloop index="idx_str" list="#qAbrvs#">      <!--- Is this correct?  I think not. --->
        <cfset output_str = Replace(output_str, "#idx_str#", )  <!--- Is this correct?  I think not. --->
        </cfloop>               <!--- What am I looping on?  What is the index? How do I do the string replacement? --->
            <!--- The chunck below is a parital listing from my Delphi Object Pascal function that does the same thing
                   I need to know how to write this part in CF9
          while not Eof do
            begin
              s := StringReplace(s, '[' +FieldByName('TBL_NM').AsString + ']', FieldByName('ABRV').AsString, [rfReplaceAll]);
              Next;
            end;
            --->
        <cfreturn output_txt>
        </cffunction>
    I'm mainly struggling with syntax here.  I know what I want to happen, I know how to make it happen in another programming language, just not CF9.  Thanks for any help you can provide.

    RedOctober57 wrote:...
    Thanks for any help you can provide.
    One:
    <cfset var output_str="#txt#">  <!--- I'm creating a local
    variable so I can manipulate the data handed in by the TXT parameter.
    Is this necessary or can I just use the txt parameter? --->
    No you do not need to create a local variable that is a copy of the arguments variable as the arguments scope is already local to the function, but you do not properly reference the arguments scope, so you leave yourself open to using a 'txt' variable in another scope.  Thus the better practice would be to reference "arguments.txt" where you need to.
    Two:
    I know what I want to happen, I know how to make it happen in another programming language, just not CF9.
    Then a better start would be to descirbe what you want to happen and give a simple example in the other programming language.  Most of us are muti-lingual and can parse out clear and clean code in just about any syntax.
    Three:
    <cfloop index="idx_str" list="#qAbrvs#">      <!--- Is this correct?  I think not. --->
    I think you want to be looping over your "qAbrvs" record set returned by your earlier query, maybe.
    <cfloop query="qAbrvs">
    Four:
    <cfset output_str = Replace(output_str, "#idx_str#", )  <!--- Is this correct?  I think not. --->
    Continuing on that assumption I would guess you want to replace each instance of the long string with the short string form that record set.
    <cfset output_str = Replace(output_str,qAbrs.TBLNM,qAbrs.ABRV,"ALL")>
    Five:
    </cfloop>               <!--- What am I looping on?  What is the index? How do I do the string replacement? --->
    If this is true, then you are looping over the record set of tablenames and abreviations that you want to replace in the string.

  • Execute the same query twice, get two different results

    I have a query that returns two different results:
    Oracle Version : 10.2.0.1.0
    I am running the following query on the Oracle server in SQL*Plus Worksheet.
    SELECT COUNT(*)
    FROM AEJOURNAL_S1
    WHERE CHAR_TIME BETWEEN TO_DATE('12-AUG-10 01:17:39 PM','DD-MON-YY HH:MI:SS AM') AND
    TO_DATE('13-AUG-10 14:17:34','DD-MON-YY HH24:MI:SS')
    AND DESC2 LIKE '%'
    AND DESC1 LIKE '%'
    AND DESC2 LIKE '%'
    AND ETYPE LIKE '%'
    AND MODULE LIKE '%'
    AND LEVELL = '11-WARNING'
    ORDER BY ORDD DESC;
    The very first time the query is run, it will return a count of 259. The next time the query is run, lets say, 10 seconds later, it will return a count of 260. The above query is exemplary of the kind of thing I'm trying to do. It seems like the more fields filtered against '%', the more random the count return becomes. Sometime you have to execute the query three or four times before it levels out to a consistent number.
    I'm using '%' as the default for various fields, because this was the easiest thing to do to support a data-driven Web interface. Maybe I have to 'dynamically' build the entire where clause, instead of just parameterizing the elements and having default '%'. Anyway, to eliminate the web interface for the purpose of troubleshooting the above query was run directly on the Oracle server.
    This query runs against a view. The view does a transpose of data from a table.
    Below is the view AEJOURNAL_S1
    SELECT
    CHAR_TIME,
    CHAR_INST,
    BATCH_ID,
    MIN(DECODE(CHAR_ID,6543,CHAR_VALUE)) AS ORDD,
    MIN(DECODE(CHAR_ID,6528,CHAR_VALUE)) AS AREAA,
    MIN(DECODE(CHAR_ID,6529,CHAR_VALUE)) AS ATT,
    COALESCE(MIN(DECODE(CHAR_ID,6534,CHAR_VALUE)),'N/A') AS CATAGORY,
    MIN(DECODE(CHAR_ID,6535,CHAR_VALUE)) AS DESC1,
    MIN(DECODE(CHAR_ID,6536,CHAR_VALUE)) AS DESC2,
    MIN(DECODE(CHAR_ID,6537,CHAR_VALUE)) AS ETYPE,
    MIN(DECODE(CHAR_ID,6538,CHAR_VALUE)) AS LEVELL,
    MIN(DECODE(CHAR_ID,6539,CHAR_VALUE)) AS MODULE,
    MIN(DECODE(CHAR_ID,6540,CHAR_VALUE)) AS MODULE_DESCRIPTION,
    MIN(DECODE(CHAR_ID,6541,CHAR_VALUE)) AS NODE,
    MIN(DECODE(CHAR_ID,6542,CHAR_VALUE)) AS STATE,
    MIN(DECODE(CHAR_ID,6533,CHAR_VALUE)) AS UNIT
    FROM CHAR_BATCH_DATA
    WHERE subbatch_id = 1774
    GROUP BY CHAR_TIME, CHAR_INST, BATCH_ID
    So... why does the query omit rows on the first execution? Is this some sort of optimizer issue. Do I need to rebuild indexes? I looked at the indexes, they are all valid.
    Thanks for looking,
    Dan

    user2188367 wrote:
    I have a query that returns two different results:
    Oracle Version : 10.2.0.1.0
    I am running the following query on the Oracle server in SQL*Plus Worksheet.
    SELECT COUNT(*)
    FROM AEJOURNAL_S1
    WHERE CHAR_TIME BETWEEN TO_DATE('12-AUG-10 01:17:39 PM','DD-MON-YY HH:MI:SS AM') AND
    TO_DATE('13-AUG-10 14:17:34','DD-MON-YY HH24:MI:SS')
    AND DESC2 LIKE '%'
    AND DESC1 LIKE '%'
    AND DESC2 LIKE '%'
    AND ETYPE LIKE '%'
    AND MODULE LIKE '%'
    AND LEVELL = '11-WARNING'
    ORDER BY ORDD DESC;
    The very first time the query is run, it will return a count of 259. The next time the query is run, lets say, 10 seconds later, it will return a count of 260. The above query is exemplary of the kind of thing I'm trying to do. It seems like the more fields filtered against '%', the more random the count return becomes. Sometime you have to execute the query three or four times before it levels out to a consistent number.
    I'm using '%' as the default for various fields, because this was the easiest thing to do to support a data-driven Web interface. Maybe I have to 'dynamically' build the entire where clause, instead of just parameterizing the elements and having default '%'. Anyway, to eliminate the web interface for the purpose of troubleshooting the above query was run directly on the Oracle server.
    This query runs against a view. The view does a transpose of data from a table.
    Below is the view AEJOURNAL_S1
    SELECT
    CHAR_TIME,
    CHAR_INST,
    BATCH_ID,
    MIN(DECODE(CHAR_ID,6543,CHAR_VALUE)) AS ORDD,
    MIN(DECODE(CHAR_ID,6528,CHAR_VALUE)) AS AREAA,
    MIN(DECODE(CHAR_ID,6529,CHAR_VALUE)) AS ATT,
    COALESCE(MIN(DECODE(CHAR_ID,6534,CHAR_VALUE)),'N/A') AS CATAGORY,
    MIN(DECODE(CHAR_ID,6535,CHAR_VALUE)) AS DESC1,
    MIN(DECODE(CHAR_ID,6536,CHAR_VALUE)) AS DESC2,
    MIN(DECODE(CHAR_ID,6537,CHAR_VALUE)) AS ETYPE,
    MIN(DECODE(CHAR_ID,6538,CHAR_VALUE)) AS LEVELL,
    MIN(DECODE(CHAR_ID,6539,CHAR_VALUE)) AS MODULE,
    MIN(DECODE(CHAR_ID,6540,CHAR_VALUE)) AS MODULE_DESCRIPTION,
    MIN(DECODE(CHAR_ID,6541,CHAR_VALUE)) AS NODE,
    MIN(DECODE(CHAR_ID,6542,CHAR_VALUE)) AS STATE,
    MIN(DECODE(CHAR_ID,6533,CHAR_VALUE)) AS UNIT
    FROM CHAR_BATCH_DATA
    WHERE subbatch_id = 1774
    GROUP BY CHAR_TIME, CHAR_INST, BATCH_ID
    So... why does the query omit rows on the first execution? Is this some sort of optimizer issue. Do I need to rebuild indexes? I looked at the indexes, they are all valid.
    Thanks for looking,
    DanIn fact you the first time you ran the query the data has been retrived from disk to memory , in the second time the data is already in memory so the respnse time should be faster ,but if you chagne any condition or column or letter case the optimizer will do the first step (data will be retrived from disk to memory )

  • Query result closing prematurely, Object closed error

    Hi guys
    I've discovered an unusual behaviour when iterating through a collection
    returned from a query.
    Scenario
    The code executes a query, iterates partially through the resultset and for
    each row calls a method which creates another object. Given that there
    could be thousands of rows, I commit the new objects every n times and then
    go back to the start of the loop which re-executes the query. This code runs
    fine when
    MAX_SCHEDULES_PER_TRANS <= com.solarmetric.kodo.DefaultFetchBatchSize.
    Problem
    If the DefaultFetchBatchSize is 10, and I commit every 11 iterations, then
    during the 3rd outer loop (ie commit has succeeded twice and the query has
    been executed 3 times) an exception is thrown when calling events.next() on
    the 10th element. The stack trace is included below, but it appears that
    the query resultset is somehow being closed early.
    If I increase the DefaultFetchBatchSize to 20, then I can also increase my
    MAX_SCHEDULES_PER_TRANS to 20 without getting the error.
    Any ideas?
    Regards
    Nathan
    Kodo 2.5.3
    SQLServer 2000
    Code Snippet
    private static final int MAX_SCHEDULES_PER_TRANS = 10;
    while (!terminated) {
    Collection col = (Collection) getQuery().execute(currentDate); //
    getQuery() method Creates query object once and reused in subsequent calls
    Iterator events = col.iterator();
    int counter = 0;
    boolean committed = false;
    while (events.hasNext() && !committed && !terminated) {
    Event event = (Event) events.next();
    event.schedule()
    counter++;
    if ((counter == MAX_SCHEDULES_PER_TRANS ) || !events.hasNext()) {
    try {
    context.getPersist().commitTransaction();
    committed = true;
    catch (OptimisticLockException e) {
    context.getPersist().rollbackTransaction();
    break;
    getQuery().close(col);
    Kodo.properties
    # Kodo JDO Properties configuration
    # DEVELOPMENT
    com.solarmetric.kodo.LicenseKey: xxxxxxx
    com.solarmetric.kodo.ee.ManagedRuntimeProperties=TransactionManagerName=Tran
    sactionFactory
    javax.jdo.PersistenceManagerFactoryClass=com.solarmetric.kodo.impl.jdbc.JDBC
    PersistenceManagerFactory
    javax.jdo.option.ConnectionDriverName=com.microsoft.jdbc.sqlserver.SQLServer
    Driver
    javax.jdo.option.ConnectionUserName=xxxxx
    javax.jdo.option.ConnectionPassword=xxxx
    javax.jdo.option.ConnectionURL=jdbc:microsoft:sqlserver://devsnetsql01:1433;
    SelectMethod=cursor;DatabaseName=devsnet09
    javax.jdo.option.RetainValues=true
    javax.jdo.option.RestoreValues=true
    javax.jdo.option.Optimistic=true
    javax.jdo.option.NontransactionalWrite=false
    javax.jdo.option.NontransactionalRead=true
    javax.jdo.option.Multithreaded=true
    javax.jdo.option.MsWait=5000
    javax.jdo.option.MinPool=0
    javax.jdo.option.MaxPool=0
    javax.jdo.option.IgnoreCache=false
    com.solarmetric.kodo.FlushBeforeQueries=with-connection
    com.solarmetric.kodo.impl.jdbc.schema.DBSequenceFactory=20
    com.solarmetric.kodo.impl.jdbc.StatementCacheMaxSize=0
    com.solarmetric.kodo.impl.jdbc.WarnOnPersistentTypeFailure=true
    com.solarmetric.kodo.impl.jdbc.SequenceFactoryClass=com.solarmetric.kodo.imp
    l.jdbc.schema.DBSequenceFactory
    com.solarmetric.kodo.impl.jdbc.FlatInheritanceMapping=true
    com.solarmetric.kodo.impl.jdbc.AutoReturnTimeout=10
    com.solarmetric.kodo.Logger=stdout
    com.solarmetric.kodo.EnableQueryExtensions=true
    com.solarmetric.kodo.DefaultFetchThreshold=30
    com.solarmetric.kodo.DefaultFetchBatchSize=10
    com.solarmetric.kodo.impl.jdbc.TransactionIsolation=READ_COMMITTED
    com.solarmetric.kodo.ee.ManagedRuntimeProperties=TransactionManagerMethod=co
    m.ibm.ejs.jts.jta.TransactionManagerFactory.getTransactionManager
    com.solarmetric.kodo.ee.ManagedRuntimeClass=com.solarmetric.kodo.ee.Invocati
    onManagedRuntime
    Stack trace
    com.solarmetric.kodo.runtime.DataStoreException: java.sql.SQLException:
    [Microsoft][SQLServer 2000 Driver for JDBC]Object has been closed.
    [code=0;state=HY000]
    NestedThrowables:
    java.sql.SQLException: [Microsoft][SQLServer 2000 Driver for JDBC]Object has
    been closed.
    at
    com.solarmetric.kodo.impl.jdbc.runtime.SQLExceptions.throwDataStore(SQLExcep
    tions.java:64)
    at
    com.solarmetric.kodo.impl.jdbc.runtime.LazyResultList.instantiateRow(LazyRes
    ultList.java:223)
    at
    com.solarmetric.kodo.impl.jdbc.runtime.LazyResultList.get(LazyResultList.jav
    a:142)
    at java.util.AbstractList$Itr.next(AbstractList.java:431)
    at
    com.solarmetric.kodo.runtime.objectprovider.ResultListIterator.next(ResultLi
    stIterator.java:49)
    at com.mallesons.servicenet.ta.shared.BOIterator.next(BOIterator.java:48)
    at
    com.mallesons.servicenet.ta.event.SchedulerThread.process(Scheduler.java:171
    at com.mallesons.servicenet.ta.event.SchedulerThread.run(Scheduler.java:122)
    NestedThrowablesStackTrace:
    java.sql.SQLException: [Microsoft][SQLServer 2000 Driver for JDBC]Object has
    been closed.
    at com.microsoft.jdbc.base.BaseExceptions.createException(Unknown
    Source)(Inlined Compiled Code)
    at com.microsoft.jdbc.base.BaseExceptions.getException(Unknown
    Source)(Inlined Compiled Code)
    at com.microsoft.jdbc.base.BaseConnection.validateClosedState(Unknown
    Source)(Inlined Compiled Code)
    at com.microsoft.jdbc.base.BaseStatement.validateClosedState(Unknown
    Source)(Inlined Compiled Code)
    at com.microsoft.jdbc.base.BaseResultSet.validateClosedState(Unknown
    Source)(Compiled Code)
    at com.microsoft.jdbc.base.BaseResultSet.getRow(Unknown Source)
    at
    com.solarmetric.datasource.ResultSetWrapper.getRow(ResultSetWrapper.java:465
    at
    com.solarmetric.kodo.impl.jdbc.runtime.LazyResultList.instantiateRow(LazyRes
    ultList.java:186)
    at
    com.solarmetric.kodo.impl.jdbc.runtime.LazyResultList.get(LazyResultList.jav
    a:142)
    at java.util.AbstractList$Itr.next(AbstractList.java:431)
    at
    com.solarmetric.kodo.runtime.objectprovider.ResultListIterator.next(ResultLi
    stIterator.java:49)
    at com.mallesons.servicenet.ta.shared.BOIterator.next(BOIterator.java:48)
    at
    com.mallesons.servicenet.ta.event.SchedulerThread.process(Scheduler.java:171
    at com.mallesons.servicenet.ta.event.SchedulerThread.run(Scheduler.java:122)

    Thanks for pointing out the optimisticLockException catch error. That would
    have been a bug but it hasn't been hit yet as I've only ever run one
    instance of this app!!
    The BOIterator is just wrapper layer and I have removed it to ensure it is
    not the culprit. Still getting the same error.
    I have done some more testing and have narrowed down the problem a little.
    It seems to only occur when I update one of the columns that is being used
    in the query. If I update a column that is not used in the query, it works
    fine.
    com.solarmetric.kodo.runtime.DataStoreException: java.sql.SQLException:
    [Microsoft][SQLServer 2000 Driver for JDBC]Object has been closed.
    [code=0;state=HY000]
    NestedThrowables:
    java.sql.SQLException: [Microsoft][SQLServer 2000 Driver for JDBC]Object has
    been closed.
    at
    com.solarmetric.kodo.impl.jdbc.runtime.SQLExceptions.throwDataStore(SQLExcep
    tions.java:64)
    at
    com.solarmetric.kodo.impl.jdbc.runtime.LazyResultList.instantiateRow(LazyRes
    ultList.java:223)
    at
    com.solarmetric.kodo.impl.jdbc.runtime.LazyResultList.get(LazyResultList.jav
    a:142)
    at java.util.AbstractList$Itr.next(AbstractList.java:431)
    at
    com.solarmetric.kodo.runtime.objectprovider.ResultListIterator.next(ResultLi
    stIterator.java:49)
    at
    com.mallesons.servicenet.ta.event.SchedulerThread.process(Scheduler.java:197
    at com.mallesons.servicenet.ta.event.SchedulerThread.run(Scheduler.java:143)
    NestedThrowablesStackTrace:
    java.sql.SQLException: [Microsoft][SQLServer 2000 Driver for JDBC]Object has
    been closed.
    at com.microsoft.jdbc.base.BaseExceptions.createException(Unknown Source)
    at com.microsoft.jdbc.base.BaseExceptions.getException(Unknown Source)
    at com.microsoft.jdbc.base.BaseConnection.validateClosedState(Unknown
    Source)
    at com.microsoft.jdbc.base.BaseStatement.validateClosedState(Unknown Source)
    at com.microsoft.jdbc.base.BaseResultSet.validateClosedState(Unknown Source)
    at com.microsoft.jdbc.base.BaseResultSet.getRow(Unknown Source)
    at
    com.solarmetric.datasource.ResultSetWrapper.getRow(ResultSetWrapper.java:478
    at
    com.solarmetric.kodo.impl.jdbc.runtime.LazyResultList.instantiateRow(LazyRes
    ultList.java:186)
    at
    com.solarmetric.kodo.impl.jdbc.runtime.LazyResultList.get(LazyResultList.jav
    a:142)
    at java.util.AbstractList$Itr.next(AbstractList.java:431)
    at
    com.solarmetric.kodo.runtime.objectprovider.ResultListIterator.next(ResultLi
    stIterator.java:49)
    at
    com.mallesons.servicenet.ta.event.SchedulerThread.process(Scheduler.java:197
    at com.mallesons.servicenet.ta.event.SchedulerThread.run(Scheduler.java:143)
    2003-10-27 15:26:40,142 DEBUG - SchedulerThread.run() - exit
    "Marc Prud'hommeaux" <[email protected]> wrote in message
    news:[email protected]...
    Nathan-
    I guess you are right that it doesn't look like the query result is
    being kept open between commits. I do notice that the block that catches
    the OptimisticLockException doesn't break out of the loop though: can
    you check to ensure that one isn't getting thrown?
    Also, what is the com.mallesons.servicenet.ta.shared.BOIterator stuff?
    If it is a wrapper layer that you have around the query execution,
    can you try removing it, just to ensure that it isn't misbehaving in
    some way?
    In article <[email protected]>, nathan boyes wrote:
    Hi Marc,
    Bug 718 certainly seems similar, however I still don't think it is
    necessarily the problem in this case. That particular bug relates to the
    resultset that was opened prior to commit, not remaining open after the
    commit. However, once I've committed, I call query.closeAll() then re
    execute the query. I would expect this to give me a new resultset that
    would not be affected by the previous commit.
    Also the CURSOR_CLOSE_ON_COMMIT is disabled by default and we haven't
    enabled it. A query against the database using one of the SQLServer
    functions confirms that this setting is false on the server.
    Predictably, instantiating the entire collection also fixes the problem.
    I
    guess this is similar to setting DefaultFetchThreshold = -1.
    Not really sure what else to try.
    "Marc Prud'hommeaux" <[email protected]> wrote in message
    news:[email protected]...
    Nathan-
    This sounds to me like
    http://bugzilla.solarmetric.com/show_bug.cgi?id=718 , even though it is
    not known to happen for SQLServer. If, before you commit, you
    instantiate the entire "col" list (with something like
    "new LinkedList (col)"), do you still see the problem?
    If this is indeed the problem, it seems that this is a server option
    that you can configure. If you disable CURSOR_CLOSE_ON_COMMIT on the
    server, do you still see the problem? See:
    http://msdn.microsoft.com/library/default.asp?url=/library/en-us/tsqlref/ts_set-set_48kz.asp
    >>>
    >>>
    >>>
    >>>
    In article <[email protected]>, nathan boyes wrote:
    I have upgraded to 2.5.5 and am now calling query.closeAll() neither
    of
    which had any effect. When I set DefaultFetchThreshold to -1, thenthe
    problem disappeared.
    For efficiency, if I am iterating through 50 objects beforecommitting
    then
    it is probably worth setting
    DefaultFetchThreshold=50
    DefaultFetchBatchSize=50
    which will run without problem.
    However, that doesn't address the underlying problem that could stilloccur
    in other parts of our system.
    "Stephen Kim" <[email protected]> wrote in message
    news:[email protected]...
    First can you upgrade to Kodo 2.5.4 and see if the problem persists?
    If so, can you for the sake of debugging try setting
    DefaultFetchThreshold to -1?
    Also you should be calling either query.closeAll () or query.close(events).
    nathan boyes wrote:
    Hi guys
    I've discovered an unusual behaviour when iterating through a
    collection
    returned from a query.
    Scenario
    The code executes a query, iterates partially through the
    resultset
    and
    for
    each row calls a method which creates another object. Given that
    there
    could be thousands of rows, I commit the new objects every n timesand
    then
    go back to the start of the loop which re-executes the query. This
    code
    runs
    fine when
    MAX_SCHEDULES_PER_TRANS <=
    com.solarmetric.kodo.DefaultFetchBatchSize.
    Problem
    If the DefaultFetchBatchSize is 10, and I commit every 11
    iterations,
    then
    during the 3rd outer loop (ie commit has succeeded twice and the
    query
    has
    been executed 3 times) an exception is thrown when calling
    events.next()
    on
    the 10th element. The stack trace is included below, but it
    appears
    that
    the query resultset is somehow being closed early.
    If I increase the DefaultFetchBatchSize to 20, then I can also
    increase
    my
    MAX_SCHEDULES_PER_TRANS to 20 without getting the error.
    Any ideas?
    Regards
    Nathan
    Kodo 2.5.3
    SQLServer 2000
    Code Snippet
    private static final int MAX_SCHEDULES_PER_TRANS = 10;
    while (!terminated) {
    Collection col = (Collection) getQuery().execute(currentDate);
    getQuery() method Creates query object once and reused insubsequent
    calls
    Iterator events = col.iterator();
    int counter = 0;
    boolean committed = false;
    while (events.hasNext() && !committed && !terminated) {
    Event event = (Event) events.next();
    event.schedule()
    counter++;
    if ((counter == MAX_SCHEDULES_PER_TRANS ) ||
    !events.hasNext()) {
    >>>>>>
    try {
    context.getPersist().commitTransaction();
    committed = true;
    catch (OptimisticLockException e) {
    context.getPersist().rollbackTransaction();
    break;
    getQuery().close(col);
    Kodo.properties
    # Kodo JDO Properties configuration
    # DEVELOPMENT
    com.solarmetric.kodo.LicenseKey: xxxxxxx
    com.solarmetric.kodo.ee.ManagedRuntimeProperties=TransactionManagerName=Tran
    sactionFactory
    javax.jdo.PersistenceManagerFactoryClass=com.solarmetric.kodo.impl.jdbc.JDBC
    PersistenceManagerFactory
    javax.jdo.option.ConnectionDriverName=com.microsoft.jdbc.sqlserver.SQLServer
    Driver
    javax.jdo.option.ConnectionUserName=xxxxx
    javax.jdo.option.ConnectionPassword=xxxx
    javax.jdo.option.ConnectionURL=jdbc:microsoft:sqlserver://devsnetsql01:1433;
    SelectMethod=cursor;DatabaseName=devsnet09
    javax.jdo.option.RetainValues=true
    javax.jdo.option.RestoreValues=true
    javax.jdo.option.Optimistic=true
    javax.jdo.option.NontransactionalWrite=false
    javax.jdo.option.NontransactionalRead=true
    javax.jdo.option.Multithreaded=true
    javax.jdo.option.MsWait=5000
    javax.jdo.option.MinPool=0
    javax.jdo.option.MaxPool=0
    javax.jdo.option.IgnoreCache=false
    com.solarmetric.kodo.FlushBeforeQueries=with-connection
    com.solarmetric.kodo.impl.jdbc.schema.DBSequenceFactory=20
    com.solarmetric.kodo.impl.jdbc.StatementCacheMaxSize=0
    com.solarmetric.kodo.impl.jdbc.WarnOnPersistentTypeFailure=true
    com.solarmetric.kodo.impl.jdbc.SequenceFactoryClass=com.solarmetric.kodo.imp
    l.jdbc.schema.DBSequenceFactory
    com.solarmetric.kodo.impl.jdbc.FlatInheritanceMapping=true
    com.solarmetric.kodo.impl.jdbc.AutoReturnTimeout=10
    com.solarmetric.kodo.Logger=stdout
    com.solarmetric.kodo.EnableQueryExtensions=true
    com.solarmetric.kodo.DefaultFetchThreshold=30
    com.solarmetric.kodo.DefaultFetchBatchSize=10
    com.solarmetric.kodo.impl.jdbc.TransactionIsolation=READ_COMMITTED
    com.solarmetric.kodo.ee.ManagedRuntimeProperties=TransactionManagerMethod=co
    m.ibm.ejs.jts.jta.TransactionManagerFactory.getTransactionManager
    com.solarmetric.kodo.ee.ManagedRuntimeClass=com.solarmetric.kodo.ee.Invocati
    onManagedRuntime
    Stack trace
    com.solarmetric.kodo.runtime.DataStoreException:java.sql.SQLException:
    [Microsoft][SQLServer 2000 Driver for JDBC]Object has been closed.
    [code=0;state=HY000]
    NestedThrowables:
    java.sql.SQLException: [Microsoft][SQLServer 2000 Driver forJDBCObject
    has
    been closed.
    at
    com.solarmetric.kodo.impl.jdbc.runtime.SQLExceptions.throwDataStore(SQLExcep
    tions.java:64)
    at
    com.solarmetric.kodo.impl.jdbc.runtime.LazyResultList.instantiateRow(LazyRes
    ultList.java:223)
    at
    com.solarmetric.kodo.impl.jdbc.runtime.LazyResultList.get(LazyResultList.jav
    a:142)
    at java.util.AbstractList$Itr.next(AbstractList.java:431)
    at
    com.solarmetric.kodo.runtime.objectprovider.ResultListIterator.next(ResultLi
    stIterator.java:49)
    atcom.mallesons.servicenet.ta.shared.BOIterator.next(BOIterator.java:48)
    >>>>>>
    at
    com.mallesons.servicenet.ta.event.SchedulerThread.process(Scheduler.java:171
    atcom.mallesons.servicenet.ta.event.SchedulerThread.run(Scheduler.java:122)
    >>>>>>
    NestedThrowablesStackTrace:
    java.sql.SQLException: [Microsoft][SQLServer 2000 Driver forJDBCObject
    has
    been closed.
    at com.microsoft.jdbc.base.BaseExceptions.createException(Unknown
    Source)(Inlined Compiled Code)
    at com.microsoft.jdbc.base.BaseExceptions.getException(Unknown
    Source)(Inlined Compiled Code)
    at
    com.microsoft.jdbc.base.BaseConnection.validateClosedState(Unknown
    Source)(Inlined Compiled Code)
    atcom.microsoft.jdbc.base.BaseStatement.validateClosedState(Unknown
    Source)(Inlined Compiled Code)
    atcom.microsoft.jdbc.base.BaseResultSet.validateClosedState(Unknown
    Source)(Compiled Code)
    at com.microsoft.jdbc.base.BaseResultSet.getRow(Unknown Source)
    at
    com.solarmetric.datasource.ResultSetWrapper.getRow(ResultSetWrapper.java:465
    at
    com.solarmetric.kodo.impl.jdbc.runtime.LazyResultList.instantiateRow(LazyRes
    ultList.java:186)
    at
    com.solarmetric.kodo.impl.jdbc.runtime.LazyResultList.get(LazyResultList.jav
    a:142)
    at java.util.AbstractList$Itr.next(AbstractList.java:431)
    at
    com.solarmetric.kodo.runtime.objectprovider.ResultListIterator.next(ResultLi
    stIterator.java:49)
    atcom.mallesons.servicenet.ta.shared.BOIterator.next(BOIterator.java:48)
    >>>>>>
    at
    com.mallesons.servicenet.ta.event.SchedulerThread.process(Scheduler.java:171
    atcom.mallesons.servicenet.ta.event.SchedulerThread.run(Scheduler.java:122)
    >>>>>>
    >>>>>>
    >>>>>
    Stephen Kim
    [email protected]
    SolarMetric, Inc.
    http://www.solarmetric.com
    Marc Prud'hommeaux [email protected]
    SolarMetric Inc. http://www.solarmetric.com
    Marc Prud'hommeaux [email protected]
    SolarMetric Inc. http://www.solarmetric.com

  • Query result Alternate row with different color   

    hello
    can anyone help me on a query result.
    i wish to have a :
    query result Alternate row with different color .
    ex.
    first row light grey
    second row darker grey.
    and repeats itself up to the last record/row.
    thanks

    <tr> <!------------------------------I replaced with
    <tr bgcolor="<cfif currentrow mod
    2>##D3D3D3<cfelse>##F5F5F5</cfif>">
    <td>
    #MYARRAY[x][2].CNO#
    </td>
    <td align="center">
    #MYARRAY[x][2].CDCDTt#
    </td>
    <td>
    #MYARRAY[x][2].PADESC#
    </td>
    <td align="center">
    #MYARRAY[x][2].INc#
    </td>
    <td align="center">
    #MYARRAY[x][2].Exp#
    </td>
    <cfset thirdArray = MYARRAY[x][3]>
    <cfif NOT arraylen(thirdArray)>
    <td> </td>
    <td> </td>
    <td> </td>
    <td> </td>
    <td> </td>
    <td> </td>
    <cfelse>
    <cfloop index="z" from="1" to="3">
    <cfif z GT arraylen(thirdArray)>
    <td> </td>
    <td> </td>
    <cfelse>
    <td>
    #thirdArray[z][2]# 
    </td>
    <td>
    #thirdArray[z][3]# 
    </td>
    </cfif>
    </cfloop>
    <cfif arrayLen(thirdArray) gt 3>
    <td nowrap="nowrap">
    <cfloop index="z" from="4" to="#arraylen(thirdArray)#">
    #thirdArray[z][2]# - #thirdArray[z][3]#<BR />
    </cfloop>
    </td>
    </cfif>
    </cfif>
    </tr>
    I got this error.
    Variable CURRENTROW is undefined

  • Difference between Replacement with a query and Replacement from Variable

    Hi Gurus
    Would explain the difference between the Replacement Path: Replacement with a query and Replacement Path: Replacement from Variable with scenarios.
    Thanks in advance,
    Aravind.S

    Hi,
    If you are using Replacement with query, the given query output will be used as input your variable. For example if you want to display the product details, which can be find our from a query.
    And Replacement with Variable will be used  when you want to give the same input  in two different fields. for example if you want to see the records from sending location and not the records from recived location at the same time. Here once you give the sending location id, it will be taken as the same input for received location.
    Regards,
    Vishnu

  • How to add different query results in Report designer

    Hi,
    I am creating a report based on 3 queries.  At the end i need to show total of 3 query results.
    How can i add results of queries in report designer??
    thanks in advance

    Hi Siva,
    As u said that u want to show the results of the queries through report designer but i think it is used for formatted reports .
    If u want to see tha results of diffrent query then u can use workbook .
    In workbook u can also analyse reports easily by applying slice and dice techniques .
    from different queries result set will be displayed on workbook with formatting  and good look and feel.
    regards,
    Supriya

  • Mapping in interconnect between different Business Objects

    I want to know how to do transformation and mapping between different business objects in interconnects.
    Always,We have a very complex SQL,when We do intergration
    with Oracle interconnect ,We use DB Adapter or Jdbc Adapter,but the complex SQL have to be excuted in the resource DB or the destination DB which may be a big pressure to them ,I think can We use different Business Objects, and do the Mappings in interconnect,so the big pressure will be on the interconnect server just like the ETL tools, But I just find that Interconnect can do tranformation and mapping in one Business Object ,how can I do? Is anyone meet this problem like me ?thanks for discussion.

    For me, Business Objects are logical groupings of business processes. For example, we have a Business Object called "Maintain_Employees". Under this we have 1 Procedure (Create_Employee) and 2 Events (Update_Employee and Delete_Employee).
    We have 1 Oracle system interfacing with 23 other legacy systems. Some of these legacy systems will be using this "Maintain_Employees" Business Object (Common View), and our main transformations will be between the Common View and the legacy Application Views.
    We are using a number of techniques to assist in "validating" data in the InterConnect. The main ones are using 'Cross Reference Tables (XREF)' and 'DatabaseOperation' transformations. By using 'Content Based Routing' we are able to send the right message to the right legacy system, and therefore do the right transformation/validation on the message payload. However, this is only a small part of a complex puzzle.
    I also have the "problem" of having "very complex SQL" on our Oracle system too. This is not unusual when using the InterConnect.
    To my mind, the InterConnect does 2 main operations. Firstly, it performs some message transformation (mapping), and secondly, it acts as a transportation engine (routing) using the adapters.
    The remainder of the effort required to create or consume the message resides with the Applications themselves. Whether it is parsing an XML CLOB payload, inserting data into staging tables, writing to log files, pre-processing data, calling API's or something else, your Application side programming and processing overhead can get large.
    The trade off it to ask the question, do I want to be able to track and manage messages from start to finish in high detail? Or can I trust that all message payload data will be consumed with no additional processing on the Application side?
    My experience has shown that the bottleneck is always at the Application side, and almost never in the InterConnect.
    The short answer to your first question is "You are right. Mappings can take place only between Application Views and Common Views only - not between Business Objects.".
    To answer your second question "Probably everyone reading this forum has this problem. The intelligence that is able to really interpret message data, validate it and process it is only found in the Application, not the InterConnect. You could, however, use the Workflow engine within OAI in order to provide additional pre-validation, human interaction and logic, but this too could be complex."
    At my current client, we are architecting an Application OAI Message handling schema. This will contain staging tables, pre-processing tables, "OAI" wrapper PL/SQL scripts, "APPS" wrapper PL/SQL scripts and Message Logging and Exception tables. Ours will be a complex set of PL/SQL processes too.
    I hope this helps, just in letting you know that you are not alone with this problem.
    I wonder if anyone else would like to share how they have architected their InterConnect and Application side mapping and transformation solutions.

  • Drag and drop between different Flash objects

    It is possible to drag & drop items between different flash objects?

    Yes, you can drag and drop anywhere on the stage.  You'll need to explain things in more detail if there's more to your question.

  • Compare DB objects between different schemas

    Hi,
    I want to compare db objects beween 2 different schemas like QA and PROD.
    Please suggest me a tool that compares table definitions,stored procedures and other db objects between different schemas.

    Did you check the above video link which is shared?
    Yes I'd agree with you and also i faced these situations to check what objects i need to export from Dev to Quality ( to analyze what are already correct or what have to be imported again)
    Generally Quality & Prod are in sync.
    Also will be waiting along with you if anyone comments on any such tool.
    Regards,
    Krishna Tangudu

Maybe you are looking for

  • Returning result set from procedure out parameter, display with anon block

    I'm trying to do something pretty simple (I think it should be simple at least). I want to use a pl/sql procedure to return a result set in an OUT parameter. If I run this code by itself (in the given anonymous block at the end, without trying to dis

  • ALE  error Function module not allowed: AFS_RETAIL_ARTMAS_IDOC_INPUT.

    Hi Gurus, By ALE config I am trying to send material from one client to another  ( MATMAS ) I have created two logical systems and assigned them to corresponding clients. Created a RFC destination and assigned logical systems. Created Distribution mo

  • Migration questions/ request for pointers

    I have two disk image backups of my system. One old, one after my system was recently corrupted, but new. I've asked around with the affected applications and tried an Archive & install keepings users profiles. That did not fix my problems. (I've tri

  • Those darn regular expressions again ...

    Greetings, I feel reluctant to ask this question because I sincerely hate regular expressions, but here is a regular expression question: Suppose I receive a byte stream from a device: after having received a, say, 'X' byte some more bytes follow, fo

  • Why isn't my library alphabetized correctly?

    So I have a lot of movie soundracks in my library, including all the Harry Potters and Lord of the Rings. It always annoyed me how when I sorted my music by album, they were always out of order, so I added the numbers 1-8 after each "Harry Potter" (E