Map CdC op transform SEQUENCE and ROW oper column properties

I am using Map_cdc_op transform, for sequence column and row operation column properties, what fields should i choose?
My source table has the following columns:
RMID - primary key int
RMName - varchar
created_date datetime
Last_updated datetime
target table also has above 4 columns plus one rmLogID column with int data type. which is key column in target.
does map_cdc_op transform also take care of the surrogate key column id in target? which is "rmlogid".
thank you very much for the helpful info.

Suneer I am using SQL serverdatabase, this table is dragged to workspace as source:
RMID - primary key int
RMName - varchar
created_date datetime
Last_updated datetime
My task is to capture data changes from this table to target table, for that i a using Map_op_CDC transform, now under the properties SEQUENCE and ROW oper column what fields should i pic.I never used this transform before.
RMid is the primary key in source, rest teh of teh fields are changeable. i track created_dt and last_updated dates for that row.
Thanks a lot for the helpful info.

Similar Messages

  • Difference between parallel sequence and parallel operation in a routing.

    Hi Experts,
    Can any one explain me with example the difference between parallel sequence and parallel operation in a routing? wHEN CAN WE USE PARALLEL OPEARTION AND PARALLEL SEQUNCE WITH COMPONENT ALLOCATION.
    Regards
    Deepak sharma

    I think u need to modify ur quest... i think u r asking about Parallel sequence and alternate seq. Below are the details from SAP site.
    A parallel sequence enables you to process several operations at the same time.
    You use an alternative sequence for example, if
    --The production flow is different for certain lot-size ranges
    For instance you can machine a work piece on conventional machine or on NC machines. A NC machine has a longer set-up time than a conventional machine. However the machining costs are considerably less. Therefore whether you use NC machines will depend on the lot size.
    ---The production flow changes under certain business conditions.
    For instance, if you have a capacity problem, you have some production steps performed externally by a vendor.

  • Transpose columns and rows / Switch columns and rows

    Hello,
    Is it possible to interchange columns and rows in order to create a left-to-right scrolling table instead of a top-to-bottom scrolling one?
    In detail:
    I have this:
          | col 1 | col 2 | col 3
    row 1 |       |       |
    row 2 |       |       |
    row 3 |       |       |
    ...   |       |       |
    ...and want to have this:
          | row 1 | row 2 | row 3 | ...
    col 1 |       |       |       |
    col 2 |       |       |       |
    col 3 |       |       |       |
    Does anyone know if this is possible with the standard table or the ALV?
    Thanks in advance & Kind regards,
    Robert

    Hi,
    So there is no easy solution by just setting a flag in the ALV config!? ...too bad!
    As for the dynamic creation:
    Would there be an issue with different data types in one "column" (former row)? As a matter of fact there probably will be a mix of character, numbers, ...
    Regards,
    Robert

  • Finding missed sequence numbers and rows from a fact table

    Finding missed sequence numbers and rows from a fact table
    Hi
    I am working on an OLAP date cube with the following schema:
    As you can see there is a fact transaction with two dimensions called cardNumber and Sequence. Card dimension contains about three million card numbers. 
    Sequence dimension contains a sequence number from 0 to 255. Fact transaction contains about 400 million transactions of those cards.
    Each transaction has a sequence number in 0 to 255 ranges. If sequence number of transactions of a card reaches to 255 the next transaction would get 0 as a sequence number.
    For example if a card has 1000 transactions then sequence numbers are as follows;
    Transaction 1 to transaction 256 with sequences from 0 to 255
    Transaction 257 to transaction 512 with sequences from 0 to 255
    Transaction 513 to transaction 768 with sequences from 0 to 255
    Transaction 769 to transaction 1000 with sequences from 0 to 231
    The problem is that:
    Sometimes there are several missed transactions. For example instead of sequence from 0 to 255, sequences are from 0 to 150 and then from 160 to 255. Here 10 transactions have been missed.
    How can I find all missed transactions of all cards with a MDX QUERY?
    I really appreciate for helps

    Thank you Liao
    I need to find missed numbers, In this scenario I want the query to tell the missed numbers are: 151,152,153,154,155,156,157,158,159
    Relative transactions are also missed, so I think it is impossible to get them by your MDX query
    Suppose this:
    date
    time
    sequence
    20140701
    23:22:00
    149
    20140701
    23:44:00
    150
    20140702
    8:30:00
    160
    20140702
    9:30:00
    161
    20140702
    11:30:00
    162
    20140702
    11:45:00
    163
    As you can see the sequence number of the last transaction at the 20140701 is 150
    We expecting that the first transaction of the next day should be 151 but it is 160. Those 10 transactions are totally missed and we just need to
    find missed sequence numbers

  • Read Attachments in Operations Mapping called by Transformation Step

    Hello,
    I would like to read the attachment of my input message. I created an UDF that assigns the attachment to to a target field, furthermore I have activated the option "Read Attachments" in operation mapping.
    But when I call the operation mapping in a transformation step in BPM, the target field is not filled. Is it not possible to read attachments in BPM?
    Thank you for your advice.

    Attachments are not available inside BPM.
    Try to do the mapping outside BPM.

  • Row to column mapping and  combining columns

    Hi
    I have one issue i want to concatenating some columns and map row as columns like below.I am giving one ex and have more than 1000 employees like this
    Example
    {Srno  Eno Ename job datestart}
    {1 01 jack------clerk------01-Jan-2008}
    { 2         01   jack-----snrclerk----01-Jun-2009}
    { 3         01    jack ---- officer------- 01-Jan-2010}
    I want to display the row to columan mapping like first row concatenated as col2 and second row for col2 column etc
    {eMpno------- col1 ------------------co12----------------------------------col3}
    {0101 jackclerk 01-Jan-2008---jacksnrclerk01-Jun-2009------01jackofficer01-Jan-2010}
    rgds
    ramya
    Edited by: user11243021 on Sep 7, 2010 5:21 AM

    with t as (
               select 1 srno,'01' eno,'jack' ename,'clerk' job,to_date('01-Jan-2008','dd-mon-yyyy') datestart from dual union all
               select 2 srno,'01' eno,'jack' ename,'snrclerk' job,to_date('01-Jun-2009','dd-mon-yyyy') datestart from dual union all
               select 3 srno,'01' eno,'jack' ename,'officer' job,to_date('01-Jan-2010','dd-mon-yyyy') datestart from dual
              ) -- end of data sample
    select  eno,
            max(
                case srno
                  when 1 then ename || ' ' || job || to_char(datestart,' dd-Mon-yyyy')
                end
               ) col1,
            max(
                case srno
                  when 2 then ename || ' ' || job || to_char(datestart,' dd-Mon-yyyy')
                end
               ) col2,
            max(
                case srno
                  when 3 then ename || ' ' || job || to_char(datestart,' dd-Mon-yyyy')
                end
               ) col3
      from  t
      group by eno
    EN COL1                      COL2                      COL3
    01 jack clerk 01-Jan-2008    jack snrclerk 01-Jun-2009 jack officer 01-Jan-2010
    SQL> SY.

  • Quick Time 7.0 Pro (7.7.3) on Win 7 Home Prem 64-bit, Pro paid for, valid registration code, Open Image Sequence (and resulting time-lapse movies) has quit working after several successful operations.  What do?  How fix? Thanks.

    Quick Time 7.0 Pro (7.7.3) on Win 7 Home Prem 64-bit, Pro paid for, valid registration code, Open Image Sequence (and resulting time-lapse movies) has quit working after several successful operations.  What do?  How fix? Thanks.

    This tip is ready for submission.
    It REPLACES the prior version in the Time Machine forum, at: http://discussions.apple.com/thread.jspa?threadID=2057525
    Additions/wording changes to items C3, C4, C5, D4
    New topics C9, C10, C11, and E4

  • Branch operation and return operation for routing in alternative sequence

    Hi all experts.
    I would like to ask which table and field store the value of branch operation and return operation for routing in the alternative sequence. I can only find them in structure (PLFLD-VORNR1 and PLFLD-VORNR2) but I want the table field of them. Would anyone please suggest a solution? Thank you.

    Hi,
    look in table AFFL, here the BKNT1 and BKNT2 refer to those vornr1 and vornr2 of the operation. For example in user-exit ZXCO1U01 you will find the values (BKNT1/2) in the sequence table, and when, for example, filling the LOIPRO01-Idoc, you can use the found VORNR for the E1AFFLL-segment.
    See coding for example on determining VORNR1/2 (sequence_table/operation_table as in user exit EXIT_SAPLCOBT_001):
    SORT operation_table BY vornr.
      LOOP AT sequence_table INTO l_sequence.
        MOVE-CORRESPONDING l_sequence TO l_affl.
    * Read the operation table to determine the vornr1 and vornr2
    *via the value of field BKNT1/2
        IF NOT l_sequence-bknt1 IS INITIAL.
          READ TABLE operation_table INTO l_operation
          WITH KEY aufpl = l_sequence-aufpl
                   aplzl = l_sequence-bknt1.
          IF sy-subrc = 0.
            l_affl-vornr1 = l_operation-vornr.
          ENDIF.
        ENDIF.
        IF NOT l_sequence-bknt2 IS INITIAL.
          READ TABLE operation_table INTO l_operation
          WITH KEY aufpl = l_sequence-aufpl
                   aplzl = l_sequence-bknt2.
          IF sy-subrc = 0.
            l_affl-vornr2 = l_operation-vornr.
          ENDIF.
        ENDIF.
    Cheers,
    S.

  • Custom Transformations Scripts: Creating sequences and synonyms

    Is there a way in Data Modeler to script the creating process of sequences and synonyms. I've already found how to create columns in a table, but is it also possible to create sequences and synonyms for tables? I've looked it up in the XML Metadata, but I didn't found where I was looking for? Can anybody help me?

    no documentation, it should be possible:
    seq = model.getStorageDesign().getSequenceSet().createSequence();
    syn =  model.getStorageDesign().getSynonymSet().createSynonym();make sure physical model is open and that model is set as physical model to your relational model. That's for Oracle
    Philip

  • Rows to columns in owb mapping

    Hi,
    I have a transformation operator in my OWB mapping where in a sequence value is generated. The sequence value is populated into two target tables.
    Now my code populates the below value 10 to two target tables which was generted by my transformation operator .
    KEY     VALUE
    1     10
    Now my requirement changed in such a way that based the key i need to generate multiple values
    i.e the output from transformation operator is
    Key     Value1     Value2
    2     10     11
    This should be populated in my target tables as
    Key     Value
    2     10
    2     11
    Please explain me how can I implement in my owb mapping.
    Many thanks in advance,
    Yeswanth

    Key is 1 and value that was generated by sequence is 10.
    in my target table both the key and the value 10 are loading now.
    After the change my target table should have
    1 10
    1 11
    The 10 and 11 are generated from the transformation operator.
    Note: Key here is not primary key nor unique.

  • How can I download the mapping from a transformation ?

    Hello gurus,
    Is there a posibility to download the mapping between the source Fields and the InfoObjects from a transformation? We have some transformations with lots of mappings and we want to have it in excel. In the transformation there is just a download option and that is in JPG (picture) and there is not what we want. In MetadataRepository I couldn't find the transformation.
    Any sugestions?
    Thanks,
    Ramona

    Hi Ramona,
    To see a transformation in Metadata Repository, you can follow this steps:
    1) Go to "Metadata Repository" in RSA1 transaction.
    2) Click on "Local objects" (the one which is placed below "Activated objects").
    3) Click on "Transformation".
    4) Press CTRL+F.
    5) Paste the transformation technical name on the search field and click "Next".  (You can copy it from RSA1 "Modeling" section).
    6) It will take you to the transformation row, where you can click on the corresponding hyperlink.
    7) Finally, you get a screen with the following details:
      a) Source Field name
      b) Description
      c) Data Type
      d) Length
      e) Rule Type
      f) Target InfoObject
      g) Description
      h) InfoObject Type
      i) Length
    8) Then, if necessary, you can copy and paste it in a Word document or Excel sheet.
    I hope this helps you.
    Regards,
    Maximiliano

  • Sequence and INSERT ALL INTO with parallelism problem

    I want to use the INSERT ALL INTO style of multiple-inserts (with 10gR2) to normalize a wide table into a skinny table. The large table has a few fk columns and 100 numeric columns. I want to insert them into another table as 100 rows in a single SQL pass.
    CREATE TABLE WIDE_SOURCE(
    FKEY_1 INTEGER,
    FKEY_2 INTEGER,
    FKEY_3 INTEGER,
    AMT001 INTEGER,
    AMT002 INTEGER,
    AMT100 INTEGER);
    CREATE TABLE SKINNY_DEST(
    PKEY INTEGER NOT NULL,
    FKEY_1 INTEGER,
    FKEY_2 INTEGER,
    FKEY_3 INTEGER,
    AMT INTEGER,
    CONSTRAINT PK_SKINNY_DEST PRIMARY KEY(PKEY)
    CREATE SEQUENCE KEY_GEN MINVALUE 1 INCREMENT BY 100 START WITH 1;
    INSERT ALL
    INTO SKINNY_DEST VALUES(KEY_GEN.NEXTVAL+001, FKEY_1, FKEY_2, FKEY_3, AMT001)
    INTO SKINNY_DEST VALUES(KEY_GEN.NEXTVAL+002, FKEY_1, FKEY_2, FKEY_3, AMT002)
    INTO SKINNY_DEST VALUES(KEY_GEN.NEXTVAL+003, FKEY_1, FKEY_2, FKEY_3, AMT003)
    INTO SKINNY_DEST VALUES(KEY_GEN.NEXTVAL+100, FKEY_1, FKEY_2, FKEY_3, AMT100)
    SELECT * FROM WIDE_SOURCE;
    With this query, each row in WIDE_SOURCE will generate 100 rows in SKINNY_DEST. I know that works. Unfortunately, it appears that the sequence isn't working the way I thought it should- with multiple nodes and parallel insert, there's some overlap in the generated key.
    Is there a better way? I've considered 100 separate sequences, but that just seems plain silly. Am I trying to have my cake (the parallelism) and eat it too (by disregarding limitations of Oracle's sequence construct)?

    Have not tested this on a RAC, but it seems to work fine - as expected without any overlaps. (version 10.2.0.1.0)
    SQL> create sequence seq_id start with 1 increment by 10 nomaxvalue nocycle;
    Sequence created.
    SQL> create table skinny ( id number, fk_key1 number );
    Table created.
    SQL> create table wide( fk_key1 number, name varchar2(32) ) parallel( degree 2 );
    Table created.
    SQL> insert into wide select object_id, object_name from all_objects where rownum < 100000;
    51853 rows created.
    SQL> commit;
    Commit complete.
    SQL> exec DBMS_STATS.Gather_Table_Stats( USER, 'WIDE' );
    PL/SQL procedure successfully completed.
    SQL> alter session enable parallel dml;
    Session altered.
    SQL> set autotrace on
    SQL> insert all
    2 into skinny values( seq_id.nextval+0, fk_key1 )
    3 into skinny values( seq_id.nextval+1, fk_key1 )
    4 into skinny values( seq_id.nextval+2, fk_key1 )
    5 into skinny values( seq_id.nextval+3, fk_key1 )
    6 into skinny values( seq_id.nextval+4, fk_key1 )
    7 into skinny values( seq_id.nextval+5, fk_key1 )
    8 into skinny values( seq_id.nextval+6, fk_key1 )
    9 into skinny values( seq_id.nextval+7, fk_key1 )
    10 into skinny values( seq_id.nextval+8, fk_key1 )
    11 into skinny values( seq_id.nextval+9, fk_key1 )
    12 select
    13 /*+ FULL(o) PARALLEL(w,2) */
    14 fk_key1
    15 from wide w
    16 /
    518530 rows created.
    Execution Plan
    Plan hash value: 1585486391
    | Id | Operation | Name | Rows | Bytes | Cost (%CPU)| Time | TQ |IN-OUT| PQ Distrib |
    | 0 | INSERT STATEMENT | | 51853 | 253K| 37 (14)| 00:00:01 | | | |
    | 1 | MULTI-TABLE INSERT | | | | | | | | |
    | 2 | INTO | SKINNY | | | | | | | |
    | 3 | INTO | SKINNY | | | | | | | |
    | 4 | INTO | SKINNY | | | | | | | |
    | 5 | INTO | SKINNY | | | | | | | |
    | 6 | INTO | SKINNY | | | | | | | |
    | 7 | INTO | SKINNY | | | | | | | |
    | 8 | INTO | SKINNY | | | | | | | |
    | 9 | INTO | SKINNY | | | | | | | |
    | 10 | INTO | SKINNY | | | | | | | |
    | 11 | INTO | SKINNY | | | | | | | |
    | 12 | SEQUENCE | SEQ_ID | | | | | | | |
    | 13 | PX COORDINATOR | | | | | | | | |
    | 14 | PX SEND QC (RANDOM)| :TQ10000 | 51853 | 253K| 37 (14)| 00:00:01 | Q1,00 | P->S | QC (RAND) |
    | 15 | PX BLOCK ITERATOR | | 51853 | 253K| 37 (14)| 00:00:01 | Q1,00 | PCWC | |
    | 16 | TABLE ACCESS FULL| WIDE | 51853 | 253K| 37 (14)| 00:00:01 | Q1,00 | PCWP | |
    Statistics
    37168 recursive calls
    26542 db block gets
    7610 consistent gets
    0 physical reads
    12696752 redo size
    498 bytes sent via SQL*Net to client
    1084 bytes received via SQL*Net from client
    6 SQL*Net roundtrips to/from client
    3 sorts (memory)
    0 sorts (disk)
    518530 rows processed
    SQL> set autotrace off
    SQL> commit;
    Commit complete.
    SQL> select
    2 id, count(*)
    3 from skinny
    4 group by id
    5 having count(*) > 1
    6 /
    no rows selected
    SQL>
    I checked the PQ V$ views and PQ slaves were used.

  • Date Field Displaying and DML Operations

    Hi all,
    I have an issue with displaying and updating date columns that I'm hoping someone can assist me with.
    I'm using APEX 3.0.1.
    I have a Form page with a number of fields sourced from one database table that are being populated by an Automatic Row Fetch On Load - After Header.
    The Item P6_MONTHFOR is stored as a Date datatype in the table and displayed on the form using the Date Picker (use Item Format Mask). I have a Format Mask set as 'MON-RR'. I want to ensure that the last day of the month is saved back to the database table so have been trying various calculation techniques to try and achieve this but am experiencing a variety of SQL errors!
    I have tried using LAST_DAY(:P6_MONTHFOR) in the Post Calculation Computation, or as a separate Computation After Submit.
    I have also tried having P6_MONTHFOR as a hidden column and using display Items and then trying Item calculations to then update the value of P6_MONTHFOR column prior to DML operations but to no avail.
    The only DML operations allowed on these rows are DELETE and UPDATE and I'm using an Automatic Row Processing (DML) On Submit - After Computations and Validations process to control these operations.
    Any help or suggestions greatly appreciated :-)
    Kind Regards,
    Gary.

    the function LAST_DAY is a date function, expecting a date as input. Since it is all web, the values of items are as string/varchar2. In order to use date-function, you have to first make it a date with to_date() with the format-mask (DD-MON-RR).
    In my opinion Dates are still tricky, it would be great if ApEx would have a DV() function, next to the V() and NV() functions, It is in ApExLib (of Patrick Wolf)
    Simon

  • Creation of sequence and trigger for each table!!!!!!!1

    Hi
    I am new to trigger and Sequence field. In one of my database we have many tables with fields for specifing ID numbers. Iam planning to insert the ID field with help of a Sequence and trigger...that trigger fires by adding the sequence value from the dual table. Now the point is here we r having around *60* table with ID field. And i am planning use the above process for each table by creating sequences and trigger for each table.
    Will this affects the performance of database.
    Is there any other option other than the above process, I mean other than creating sequences and trigger for each table.
    PLzz help to resolve this issuee......
    Shiyas
    Edited by: user13170361 on Jun 7, 2010 12:37 AM

    Tiger, I didn't mind about your comment, but the point is try to use
    select NVL(max(a) + 1,1) into i from p1_temp;This line in your trigger code and see what is happening. The problem is with your trigger. You are using group by function and you will not get no_data_found !
    For more help, this is some modification of your code.
    SQL> create table p1_temp (a number(10) primary key, b number(10));
    Table created.
    SQL> create or replace trigger trg_p1_temp
      2  before insert on p1_temp for each row
      3  declare
      4  i number(10);
      5  begin
      6  begin
      7  select NVL(max(a) + 1,1) into i from p1_temp;
      8  exception
      9  when no_data_found then
    10  i := 1;
    11  end;
    12  :new.a := i;
    13  end;
    14  /
    Trigger created.
    SQL> insert into p1_temp(b) values (1);
    1 row created.
    SQL> insert into p1_temp(b) values (2);
    1 row created.
    SQL> commit;
    Commit complete.
    SQL> select * from p1_temp;
             A          B
             1          1
             2          2
    SQL> Edited by: Saubhik on Jun 7, 2010 2:30 AM

  • Name and Address operator's packge do not have any values

    Hi All,
    I am using the Name and Address operator, but I do not have/want the 3rd party vendor's connector/tool/data library to work with my OWB.
    I developed the map and deployed that map into my target.
    Package got generated of name NAME_ADDRESS
    In this package the entry point is the MAIN function, in the main function taking one parameter p_env.
    When I execute the Package from SQL*Plus I am getting the value of p_env, which is wrong, p_env can not be NULL/0 at any case.
    All I want is, how to pass/populate the value into p_env parameter, there will be more than one value into that parameter, param_value varchar2(30)and param_name varchar2(4000).
    But I know nothing about what are the values may be in , when I execute the package...
    ALSO I AM ATTACHING SOME MORE DETAILS....
    create or replace package
    FUNCTION Main(p_env IN WB_RT_MAPAUDIT.WB_RT_NAME_VALUES) RETURN NUMBER IS
    end package
    In this main function, the p_env is getting Initialize, it calls the initialize
    procedure, there directly getting count on this parameter.
    my question is at what point of time these valiables get initialize and what
    are the exact details for the followings
    TYPE WB_RT_NAME_VALUE IS RECORD
    param_name varchar2(30),
    param_value varchar2(4000)
    TYPE WB_RT_NAME_VALUES IS TABLE OF
    WB_RT_NAME_VALUE
    INDEX BY BINARY_INTEGER;
    What are these two variables (param_name and param_value) implies in wb_rt_mapaudit package
    ANY HELP GREATLY APPRECIATED
    THANKS FOR YOUR TIME
    Gladson
    [email protected]

    Gladson,
    Page 8-73 of the Warehouse Builder user's guide explicitly highlights the following section:
    Note: Taking advantage of the Name and Address operator requires separate licensing and additional installation steps. Refer to the Oracle9i Warehouse Builder Installation and Configuration Guide for more information.
    The functionality that does the parsing and standardization is not part of Warehouse Builder, but part of the package that the software partner brings in. If you think about that, it actually makes sense. Would you expect Warehouse Builder to contain the code to perform this action for all countries across the globe, in all different languages and the like? The data vendors, who are the experts in the different areas, know this information.
    So, if all you want to do, is parse and standardize an address, then yes, you may be better of going directly to a partner (having said that, you may get a discount if you are a Warehouse Builder customer, but that I do not know for sure).
    However, what does Warehouse Builder add:
    - Data object design (maybe your target tables don't exist) and deployment
    - ETL design and code generation
    - Process flow creation and deployment
    - Execution management (automatic logging etc)
    - Integration with Discoverer
    - Metadata management, such as impact analysis, version management, multi language support...
    - Much, much more.
    Specific to name and address (or say data quality): if you include the name and address operator on a Warehouse Builder mapping, then you can plug in any partner library you want, and in case you find out that another partner delivers better results, then go with the other partner. At the end of the day, you abstracted the implementation from the design. That makes you much more flexible in your choice of a data vendor. Besides, you get impact analysis, lineage and lots of other metadata management features for free. On top of that, if you want to identify duplicates, perform householding and the like, then there is match and merge (which yes, is part of the tool and not a third-party implementation) in order to define the rules for matching and merging.
    Again, if you don't need any of that, then perhaps you don't want the overhead of having all this and you go directly to a data vendor.
    Hope this helps,
    Mark.

Maybe you are looking for

  • Not able to ssh out from a router

    Hi All: I notices i can't not ssh out to another device from my router. SSH in from my client (putty) works fine. and not access-list attached to vty with out direction. transport output all I found this problem happened on my ASR1002 with "asr1000rp

  • Strategy 50 for AFS material

    Hi,    In my current project we use AFS material, For a finished material we use strategy 50, after creating planned independent requirement, without running MRP, in stock requirements we can able to see the planned orders. How is this possible?. Als

  • Sync deleted my free christmas apps

    I downloaded the free games from the 12 days of Christmas app directly to my iPhone 3G and the first time I tried to sync with iTunes it told me I only had 4gb memory and had overloaded my phone. It then deleted the 3 free games and a separate free a

  • Netbeans, Matisse, Swing, JDK5 & JDK6

    This is more a related to a combination of Netbeans, the JRE and Swing. I'm developing a Swing application in Netbeans (currently NB 6 M10) with JDK/JRE6. I'm using Matisse so there is quite a bit of code automatically generated for Swing components

  • Are you going to support Bitcoin/cryptocurrencies in Firefox

    What support are you going to give to Bitcoin/cryptocurrencies.