How to make test data anonymus

Hi all,
we are planning to copy our productive system to our test system. Therefore it is necessary to make all HR and customer data anonymus.
Is there a possibility within the standard (maybe a report) to make these data anonymus.
If there is no standard report, has anybody of you already written such a report and can tell me how is the best way to make data dynamically anonymus?
Thanks in advance for your help.
Best regards
Tom

Hi,
Could you please elaborate it more?
rgds

Similar Messages

  • How to make a data base connexion in java?

    I have an oracle 7 data base that i want to connect how to make a data base connexion in java?

    J
    D
    B
    C
    Search these forums or follow the JDBC tutorial

  • How to make the date to be entered as dd-mon-rrrr  using form personalizati

    Hi friends,
    how to make the date entered to be dd-mon-rrrr format.....
    the date should be entered only in this format...or esle it should not accept the date
    how to do this with form personalization.....
    thanks

    1) Instead of personalization, set ICX: Date format mask profile appropriately using sysadmin responsibility at the appropriate level.
    2) If you have to use this new format just for one form, then use personalization to set the profile option in the when-new-form-instance. But unsetting it will be a problem.
    I recommend option 1 above.
    Hope this answers your question,
    Sandeep Gandhi

  • How to generate test data for all the tables in oracle

    I am planning to use plsql to generate the test data in all the tables in schema, schema name is given as input parameters, min records in master table, min records in child table. data should be consistent in the columns which are used for constraints i.e. using same column value..
    planning to implement something like
    execute sp_schema_data_gen (schemaname, minrecinmstrtbl, minrecsforchildtable);
    schemaname = owner,
    minrecinmstrtbl= minimum records to insert into each parent table,
    minrecsforchildtable = minimum records to enter into each child table of a each master table;
    all_tables where owner= schemaname;
    all_tab_columns and all_constrains - where owner =schemaname;
    using dbms_random pkg.
    is anyone have better idea to do this.. is this functionality already there in oracle db?

    Ah, damorgan, data, test data, metadata and table-driven processes. Love the stuff!
    There are two approaches you can take with this. I'll mention both and then ask which
    one you think you would find most useful for your requirements.
    One approach I would call the generic bottom-up approach which is the one I think you
    are referring to.
    This system is a generic test data generator. It isn't designed to generate data for any
    particular existing table or application but is the general case solution.
    Building on damorgan's advice define the basic hierarchy: table collection, tables, data; so start at the data level.
    1. Identify/document the data types that you need to support. Start small (NUMBER, VARCHAR2, DATE) and add as you go along
    2. For each data type identify the functionality and attributes that you need. For instance for VARCHAR2
    a. min length - the minimum length to generate
    b. max length - the maximum length
    c. prefix - a prefix for the generated data; e.g. for an address field you might want a 'add1' prefix
    d. suffix - a suffix for the generated data; see prefix
    e. whether to generate NULLs
    3. For NUMBER you will probably want at least precision and scale but might want minimum and maximum values or even min/max precision,
    min/max scale.
    4. store the attribute combinations in Oracle tables
    5. build functionality for each data type that can create the range and type of data that you need. These functions should take parameters that can be used to control the attributes and the amount of data generated.
    6. At the table level you will need business rules that control how the different columns of the table relate to each other. For example, for ADDRESS information your business rule might be that ADDRESS1, CITY, STATE, ZIP are required and ADDRESS2 is optional.
    7. Add table-level processes, driven by the saved metadata, that can generate data at the record level by leveraging the data type functionality you have built previously.
    8. Then add the metadata, business rules and functionality to control the TABLE-TO-TABLE relationships; that is, the data model. You need the same DETPNO values in the SCOTT.EMP table that exist in the SCOTT.DEPT table.
    The second approach I have used more often. I would it call the top-down approach and I use
    it when test data is needed for an existing system. The main use case here is to avoid
    having to copy production data to QA, TEST or DEV environments.
    QA people want to test with data that they are familiar with: names, companies, code values.
    I've found they aren't often fond of random character strings for names of things.
    The second approach I use for mature systems where there is already plenty of data to choose from.
    It involves selecting subsets of data from each of the existing tables and saving that data in a
    set of test tables. This data can then be used for regression testing and for automated unit testing of
    existing functionality and functionality that is being developed.
    QA can use data they are already familiar with and can test the application (GUI?) interface on that
    data to see if they get the expected changes.
    For each table to be tested (e.g. DEPT) I create two test system tables. A BEFORE table and an EXPECTED table.
    1. DEPT_TEST_BEFORE
         This table has all EMP table columns and a TEST_CASE column.
         It holds EMP-image rows for each test case that show the row as it should look BEFORE the
         test for that test case is performed.
         CREATE TABLE DEPT_TEST_BEFORE
         TESTCASE NUMBER,
         DEPTNO NUMBER(2),
         DNAME VARCHAR2(14 BYTE),
         LOC VARCHAR2(13 BYTE)
    2. DEPT_TEST_EXPECTED
         This table also has all EMP table columns and a TEST_CASE column.
         It holds EMP-image rows for each test case that show the row as it should look AFTER the
         test for that test case is performed.
    Each of these tables are a mirror image of the actual application table with one new column
    added that contains a value representing the TESTCASE_NUMBER.
    To create test case #3 identify or create the DEPT records you want to use for test case #3.
    Insert these records into DEPT_TEST_BEFORE:
         INSERT INTO DEPT_TEST_BEFORE
         SELECT 3, D.* FROM DEPT D where DEPNO = 20
    Insert records for test case #3 into DEPT_TEST_EXPECTED that show the rows as they should
    look after test #3 is run. For example, if test #3 creates one new record add all the
    records fro the BEFORE data set and add a new one for the new record.
    When you want to run TESTCASE_ONE the process is basically (ignore for this illustration that
    there is a foreign key betwee DEPT and EMP):
    1. delete the records from SCOTT.DEPT that correspond to test case #3 DEPT records.
              DELETE FROM DEPT
              WHERE DEPTNO IN (SELECT DEPTNO FROM DEPT_TEST_BEFORE WHERE TESTCASE = 3);
    2. insert the test data set records for SCOTT.DEPT for test case #3.
              INSERT INTO DEPT
              SELECT DEPTNO, DNAME, LOC FROM DEPT_TEST_BEFORE WHERE TESTCASE = 3;
    3 perform the test.
    4. compare the actual results with the expected results.
         This is done by a function that compares the records in DEPT with the records
         in DEPT_TEST_EXPECTED for test #3.
         I usually store these results in yet another table or just report them out.
    5. Report out the differences.
    This second approach uses data the users (QA) are already familiar with, is scaleable and
    is easy to add new data that meets business requirements.
    It is also easy to automatically generate the necessary tables and test setup/breakdown
    using a table-driven metadata approach. Adding a new test table is as easy as calling
    a stored procedure; the procedure can generate the DDL or create the actual tables needed
    for the BEFORE and AFTER snapshots.
    The main disadvantage is that existing data will almost never cover the corner cases.
    But you can add data for these. By corner cases I mean data that defines the limits
    for a data type: a VARCHAR2(30) name field should have at least one test record that
    has a name that is 30 characters long.
    Which of these approaches makes the most sense for you?

  • How to make the date disappear on the blog summary page?

    Hi,
    I have read thru the message on this forum and can't seem to figure out how to hide or make the date disappear on the blog summary page.  I can do it on the blog entry pages by turning the color opacity to 0. 
    I have included my page here:http://www.quantumhealthcentre.com/www.quantumhealthcentre.com/Services/Services .html
    Also, if anyone knows how to delete the archive page or inactivate it on my website, please let me know.
    Any advice would be appreciated.
    Thanks,
    Peggy

    http://www.wyodor.net/_Movies/HideDateBlogEntry.mov
    http://www.wyodor.net/_Movies/HideDateBlogArchive.mov

  • How to make the date format in ipod the same as in itunes?

    When I look at my podcasts in itunes, the dates come up as dd/mm/yyyy (which is correct), however, when I look at my podcasts on the ipod, the date comes up as mm/dd, which is very confusing.
    I cannot figure out how to make it so that the date format is the same.

    When I look at my podcasts in itunes, the dates come up as dd/mm/yyyy (which is correct), however, when I look at my podcasts on the ipod, the date comes up as mm/dd, which is very confusing.
    I cannot figure out how to make it so that the date format is the same.

  • How to generate test data

    Hi
    I need to generate some test data and would like to know if there are any built functions in SQL Server 2012 that let you do that apart from using loops. In PostgreSQL I use different built in functions to achieve that for example the following statement
    creates the test data for me, if any one can show me how I can do that using T-SQL
    CREATE TABLE domain AS SELECT generate_series(1,100000) AS domain_id, substr('abcdefghijklmnopqrstuvwxyz',1, (random()*26)::integer) || '.com' AS domain_name;
    And also I would like to know how do you get help on syntax in T-SQL, for example in PostgreSQL the following commands show you help etc so are there any similar commands in T-SQL ?
    \h ALTER TABLE   ( Will show you full syntax for ALTER TABLE )
    \dt  ( Will detail all the tables in the current Schema )
    Thanks a lot
    Rgds
    T

    The two T-SQL suggestions are great and probably the simple answer you're looking for.  I wrote a blog article about using the VS data generation feature, as suggested by Chuck, to do some nice stuff here if you want to check that out:
    http://blogs.msdn.com/b/samlester/archive/2012/08/04/creating-complex-test-databases-part-2-creating-a-database-with-1-billion-random-rows.aspx
    Thanks,
    Sam Lester (MSFT)
    My Blog
    This posting is provided "AS IS" with no warranties, and confers no rights. Please remember to click
    "Mark as Answer" and
    "Vote as Helpful" on posts that help you. This can be beneficial to other community members reading the thread.

  • How to add Test data for a function module

    Hi experts,
    i want to add test data for a function module . i don't know how to proceed on it . please help me...
    with regards,
    James...
    Valuable answers will be rewarded...

    Hi,
    - Go to SE37 and execute your FM
    - Enter the data you want to pass to FM
    - Hit 'Save' button. Enter the meaningful name for your test scenario and save. That's it you have saved the test data.
    You can also enter some other test data and save that also. In short, you can save multiple test scenario. Also, you can give multiple test scenario a same name and they do not overwrite each other ( but normally you give different name to differentiate them)
    Next time you come to execute the FM again, hit the "test data" button and it will show you all the test scenario you have stored before. Select the one you want to use and it will load the data in FM parameters.
    Let me know if you need any other information.
    Regards,
    RS

  • How to create test data in SHDB transaction ?

    Hi,
       In SHDB transaction I have captured the recording of AS01 transaction. Now there is a tab called "test data" in SHDB transaction. I clicked on that and gave a text document name called "c:\asset.txt" . Now it is showing status like "test data has already been saved into c:\asset.txt". Now how should I create test data for this?
    Please reply soon...............

    Hi,
        Can anyone pls explain what the tab "Test data" will function in SHDB transaction?

  • How to make Start Date as optional parameter in Crystal reports 2011

    Hi All,
    I have existing Start Date and End Date parameter how to make them as optional parameters.
    Thanks for your help in advance!!.

    Hi Mnnica,
    There is an option 'Optional Prompt' under value options, make that as true.  Do the same for both startdate and enddate then go in Record Selection Formula and write below formula :
    If Not (hasvalue({?Startdate}) then True else {?Startdate} >= {databasedatefield}
    and
    If Not (hasvalue({?Enddate}) then True else {?Enddate} <= {databasedatefield}
    Check below screen capture for Optional Prompt options :
    -Sastry

  • How to make CRM data sources for delta capability?

    Hi All,
    I know that the CRM data sources 0CRM_SALES_ACT_1, 0CRM_QUOTATION_I, 0CRM_SALES_ORDER_I have delta capability.
    At present we are doing full loads daily but it has got some performance issues.So we are thinking to make these as delta loads.
    Already some one has made a trail to make delta loads but not successful for some unknown reasons.
    Could you please suggest me the steps to make these data sources delta capable both in CRM and BW systems?
    As this is urgent...Please suggest ASAP.

    Hi
    Go thorugh Below the note: 692195
    Summary
    Symptom
    There may be problems or issues related to data tranfer from CRM to BW.
    Other terms
    CRM-BW extraction,upload,initial,delta,full upload ,Sales Analytics,
    Reason and Prerequisites
    There could be errors in customization or program errors due to which
    data may not be transferred or incorrectly transferred to BW.
    Solutions
    Question 1 : The Extraction from CRM to BW takes a very long time. What can be done? (Performance Issues)
    Suggestion 1: Please implement notes  653645 (Collective note) and
    639072(Parallel  Processing).
    The performance could be slow because of the wrong control parameters
    used for packaging.
    You can change the package size for the data extraction.
    Also note that changing the package size in the transaction SBIW
    would imply a change for all the extractors. Instead, you could
    follow the path in the bw system.
    Infopackage (scheduler)    > Menu 'Scheduler'   > 'DataS. default data
    transfer'   > maintain the value as 1500 or 1000(This value is variable)
    The package size depends on the Resources available at the customer side
    (The no of parallel processes that could be assigned =
    1.5 times the no of CPU's available approx.)
    Question 2 : On executing transaction RSA3 I get records but I find 0
    records when I load data from BW request.(No Data Available)
    Suggestion 2: First check if there are any entries in the table
    CRMD_ORDER_INDEX.Only if there are entries in this table you can
    extract records.
    If this is not the case then,
    It is possible that the user does not have sufficient
    authorities for extraction of  the relevant objects.
    Additionally, please review and implement the following notes
    615670
    161570
    150315
    618953
    If you are in the release 4.0 then
    To do BW extraction with the user please see that the following
    authorization object exists(display mode is enough):
    CRM_ACT,CRM_OPP, CRM_LEAD, CRM_SAO, CRM_SEO, CRM_CO_SE,CRM_CO_SC
    CRM_CO_SA, CRM_CON_SE, CRM_CMP, CRM_ORD_OP,CRM_ORD_LP,CRM_ORD_PR
    CRM_ORD_OE, CRM_CO_PU, CRM_CO_PD, CRM_ORD_PO
    (all these objects are linked to transaction crmd_order).
    Question 3 : The Deltas for my data source are not extracted . What can I do?
    Suggestion 3: Please check the following.
    Please Check if the services have been generated in transaction GNRWB.
    If they are not active(not marked 'X' before their names) then activate
    the services following the steps here.
    Go to transaction GNRWB
    Select BUS_TRANS_MSG
    Select (on the right, the services) : BWA_DELTA3, BWA_FILL, BWA_queue
    Press Generate.
    Also check  for the following:
    1. The delta should have been initialized successfully.
    2. Confirm that all Bdocs of type BUS_TRANS_MSG
       are processed with success in SMW01.
    3. If there are queues in SMQ1 with erroneous status then activate
       these queues.
       In Transaction SMQ1 if there are Queues existing with
       names beginning with CRM_BWAn (n is number) then
       activate these queues in the same transaction.
    4.a)If required activate the datasource
        Go to transaction BWA5   > select the required datasource and
        activate.
    4 b) The Delta may not be active ,activate the delta in BWA7 by
      selecting the name of the datsource and pressing the candle icon for
      'activate delta'.
    5. In BW system
       Go to transaction RSA1   > modeling   > infosources   > select the
       infosource   > right mouse click on the selected
       infosource   > choose option replicate datasource
        Activate the infosource.
    6. Go to the scheduler for the infosource   > select delta  in the
        update  >choose the option PSA only (in the Processing tab)
        > start immediately
    Check the entry in the RSA7 in the OLTP(CRM system)
    Question 4: How can I extract the fields, which are not provided in the standard  data source extraction .
    Suggestion 4: Follow the steps mentioned below.
    1. Enhance Extract Structure with the required fields. (Create & include
       an append structure to the extract structure via transaction RSA6).
    2.a) Release the fields of the append  for usage. (  To do this, double
       click on the Datasource and remove the flags in the column 'Hide
       Field' for all fields of Append. )
    2.b)If the new fields cannot be seen in the extract structure of
        the transaction BWA1 then change and save the datasource, and then
        activate it in RSA6.
    3. Define your mappings in BADI (CRM_BWA_MFLOW) to fill these fields.
    Goto SPRO .
    Follow the path ->
    SAP Implementation guide ->Implementation with other mySAP components ->
    Data transfer to the Business Information Warehouse->
    Settings for the application specific datasources (CRM)->
    Settings for BW adapter->
    Badi :BW adapter :Enhancement of datasources in messaging flow.
    4. Replicate the new Datasource to BW.
    5. Expand the Communication Structure in BW.
    6. Maintain transfer Rules for the new Datasource.
    7. Activate the trasfer rules and perform the upload.
    Question 5:  I am unable to extract  user status correctly.What should I
    do?
    Suggestion 5 :Check the following notes
    531875
    616062
    713458
    700714
    765281
    Question 6 : What can I  do when the activity/Opportunity/Complaint
    reasons(Code,CodeGruppe,Katalogart)  are not extracted.
    Suggestion 6 :Check the following notes
    481686
    516820
    603609
    617411
    711146
    Question 7:Deleted opportunities are not reflected in BW.
    Suggestion 7: Check the note 706327.
    Question 8: How do you  activate the metadata?
    Suggestion 8:CRM BW adapter meta data has to be activated first before
    it is available in the system. You can use Transaction BWA5 to copy the
    meta data for selected DataSources. You can reach the transaction via
    the IMG maintenance 'SAP Reference IMG -> Settings for SAP Business
    Information Warehouse -> Activate BW Adapter Meta Data'. For more
    information, see the documentation on the IMG activity 'Activate BW
    Adapter Meta Data'.  (Note 432485)
    Question 9: I donot get any records for the delta upload of my attribute
    datasource(s).What is to be done?
    Suggestion 9: In case of attribute datasources, it is possible that the
    entry for the GUID is missing in the table SMOXAFLD.
    If , for example the datasource 0CRM_OPPT_ATTR is not giving deltas
    then you can follow the steps:
    1)If Delta process is active for the attribute datasource e.g.
         0CRM_OPPT_ATTR,  then stop the delta process in the BW
         system 
    2) In the CRM system, Make the entry in the table SMOXAFLD
       for the datasource with the Key
       as  0CRM_OPPT_ATTR     GUID 3) Save the entry.Activate the datasource                           4) Check that the above entry is replicated in smoxafld_s also    after this.               5) Create a transport request manually for the following    object             R3TR   SMO4   0CRM_OPPT_ATTR    
    Question 10: What do the status BWSTONESYS0 , BWSTONEUSS0,BWSTTECSYS0 and various other BW status mean ?
    Suggestion 10: The BW status are used to extract system and user defined
    status.
    The BW status are defined in the customization settings in SPRO.
    Check for-> Status Concept for BP/Product/CRM objects
    Here goto-> Process user status You will find the documentation attached here for the user status. Going inside the transaction you will find the status groups USS0, ZIOP,ZMOP etc. The names of the various BW status are derived from this For ex. BW + ST+ One + USS0 gives the name of the field BWSTONEUSS0 or BWST + ONE + ZMOP = BWSTONEZMOP (Master opportunity values) (which means BWST + (status group name) + status object group name) Double clicking on any of the object groups will take you to the values that these status can have . For ex. BWSTONEUSS0 in your system can have values E001 ,E002,E003,E007 which will be shown in RSA3 as BW status values 1,2,3,7, respectively.
    Similarly we have Goto -> Process system status (in SPRO). Here you can get the values for the system status in exactly the same way as BWSTONESYS0(Lifecycle status) , BWSTTECSYS2 (Error) etc. In RSA3 you get the names as BW status, To know which corresponds to which status here,Gotothe record list in RSA3 . Here goto Settings - Layout -Current . Right click -> Press Show technical field names . You will be able to see the BW status names and will be able to adjust the layout accordingly.
    Thank you,
    DST

  • How to insert test data of 10,000 records into emp table

    Hi I'm new to oracle can anyone please help me in writing a program so that i can insert test data into emp table

    Hi,
    user11202607 wrote:
    thanks sanjay , frank . But how can i insert only 4 deptno's randomly and how can i insert only 10 managers randomly ,
    Sorry to pull Your legs and thanks for bearing my question. I want to insert into emp table where it has the empno, ename, sal, job, hiredate, mgr and deptnoThis should give you some ideas:
    INSERT INTO emp (empno, ename, sal, job, hiredate, mgr, deptno)
    SELECT  LEVEL                         -- empno
    ,     dbms_random.string ('U', 4)          -- ename
    ,     ROUND ( dbms_random.value (100, 5000)
               , -2
               )                         -- sal
    ,     CASE 
               WHEN  LEVEL =  1              THEN  'PRESIDENT'
               WHEN  LEVEL <= 4            THEN  'MANAGER'     -- Change to 11 after testing
               WHEN  dbms_random.value < .5  THEN  'ANALYST'
               WHEN  dbms_random.value < .5  THEN  'CLERK'
                                                 ELSE  'SALESMAN'
         END                         -- job
    ,     TRUNC ( SYSDATE
               - dbms_random.value (0, 3650)
               )                         -- hiredate
    ,     CASE
             WHEN  LEVEL > 1
             THEN  TRUNC (dbms_random.value (1, LEVEL))
         END                         -- mgr
    ,     TRUNC (dbms_random.value (1, 5))     -- deptno
    FROM     dual
    CONNECT BY     LEVEL <= 10                         -- Change to 10000 after testing
    ;The interesting part (to me, at least) is mgr. What I've done above is guarantee that the mgr-empno relationship reflects a tree, with the 'PRESIDENT' at its sole root. The tree can be any number of levels deep.
    Sample results:
    EMPNO ENAME        SAL JOB        HIREDATE  MGR DEPTNO
        1 GDMT        2800 PRESIDENT  30-AUG-04          2
        2 CVQX         400 MANAGER    24-MAY-06   1      2
        3 QXJD        1300 MANAGER    17-JUN-05   1      4
        4 LWCK        4800 MANAGER    15-JUN-06   2      2
        5 VDKI        3700 CLERK      08-SEP-01   4      2
        6 FKZS        2600 CLERK      18-DEC-06   4      1
        7 SAKB         700 ANALYST    30-JUN-00   5      4
        8 DVYY         300 ANALYST    22-SEP-01   2      1
        9 CLEO        2700 ANALYST    27-MAY-08   5      4
       10 RDVQ        3400 ANALYST    14-DEC-08   5      4For details on the built-in packages (such as dbms_random) see the [Parckages and Types manual|http://download.oracle.com/docs/cd/B28359_01/appdev.111/b28419/d_random.htm#i998925].

  • How to install test data on iPhone

    I have some test data which I have been using in the simulator by copy it to the applications private Documents folder. I can then open this file in my application running on the simulator and use to verify results etc.
    How can I install this test data file on a physical iPhone and perform a similar function. I can see, using the Organiser, how to get data off the iPhone to the Mac but the other way around.
    I am missing something obvious?
    Thanks, Dave.

    Place it into the Resources section of your project then at runtime in your application delegate you can do something like the following (assumes your data is in a file called testfile.dat):
    NSBundle *thisBundle = [NSBundle bundleForClass:[self class]];
    NSString *fileToCopyPath = [thisBundle pathForResource: @"testfile" ofType:@".dat"];
    NSString *path = [resourceDir stringByAppendingPathComponent: @"testfile.dat"];
    NSError *error = nil;
    // replace old test data with new
    [[NSFileManager defaultManager] removeItemAtPath: path error: &error];
    [[NSFileManager defaultManager] copyItemAtPath: fileToCopyPath toPath: path error: &error];

  • How to creat Test Data Directory

    Hi Experts,
    I want to create a TEST DATA DIRECTORY for the following function module CLOI_CHANGES_UPL_31. How to create this?
    Actually now i have some process order numbers. Using this how to create the test data directory??
    Plz guide this.
    Point will be sure.
    Mohana

    Hi Mohana,
          Test data directory is similar to variants in reports.
          After executing the function module give some input and click the save button . It will prompt a name . Give the name and click enter. So this test data is saved in the test data directory.
       So that you need not not give the test input again and again. you can simply click the test dat directory and select your test data .
    Regards,
    Charumathi.B

  • How to make Filemaker data indexed by SES?

    We want to make lots of our data in Filemaker indexed by SES(Oracle Secure Enterprise Search). But it seems that oracle doesn't provide Filemaker option. How can I make the data indexed?
    Message was edited by:
    yoohan

    We want to make lots of our data in Filemaker indexed
    by SES(Oracle Secure Enterprise Search). But it seems
    that oracle doesn't provide Filemaker option. How can
    I make the data indexed?
    Is it a difficult question? I think it isn't that difficult to write a adaptor to extract the text from filemaker data record.
    Message was edited by:
    yoohan

Maybe you are looking for

  • Discover Plus - Export to Text Tab delimited is not exporting all the rows

    Hi gurus, I am trying to export a large data report which has 1 million plus rows to text tab delimited. The export takes 9 plus hours to export and the data is not more than 100000. My question is 1. How can I make the discoverer to export it quicke

  • Using Java 6 classes with Java 5

    I ran into a little problem with the portability of Java (ironic eh?) across computers. The program I wrote on my home computer uses GroupLayout for the GUI. However, much to my surprise, the computers at my school only had Java 5 installed so the pr

  • Why is the movie too bright on TV but OK on monitor?

    Hi - I don't know whether the problem is in AE or Encore so I hope I'm in the right forum. I built the 6:00 animation in AE and rendered it as a QT mov with no compression.  The subject is filling out forms and mailing them to the correct office, so

  • Urgent Doubt - Custom AccessGate way

    I am creating a custom access gateway using Oracle Access server SDK 10.1.4. Our OAM policy is configured to return the couple of variables after successful authentication and authorization. eg: ObUserSession session = new ObUserSession(obssoCookie);

  • Can't delete Previous System File

    Hi, I did an archive and install and was left with a Previous System folder that is about 13G. I used "whatsize" app to move it to the trash where I then did a Secure empty trash. Trash can is empty, but when I run Whatsize again, the same previous s