DML malfunction after application and table export

Hello guys,
I need ur help:
I try to deploy my application to another PC, I have faced some problems:
1)I try to build again all the tables,trigger ans sequence manually by pasting the SQL command from my SQL
script.
Is ithere an easier way to do that??
Can i include those necessary components such as tables,
sequences,processes in an installation package along with the application??
I am using Application Express 2.1.0.00.39 , hence there is no supporting objects script like APEX 3.0 and
above.
2) Also, when i am able to execute all the SQL , the generated table (with sequence and trigger) no longer able
to do the data manipulation job again.
It keeps stating "cannot insert null" error although i have completed all the field ..
Can anybody help me??
Thanks and best regards

keepng audit trail is very important.
secondlly users are known...
they are around 30 and i beleive they may increase 5 users per year.
its in internal LAN,, and just replace the desktop application.
I do not understand by connection pooling problem.
kindly help out with a few more suggestions.
Do you feel it ok to have a single DB users, and all other application users.
From DBA point of viw is there something wrong in it.
IF we make application users, then application code will get user name and insert into history table.
if we make db users then Database trigger can do the same work as well.
hope you are getting me
thanks alot for your kind suggestion.
plz reply early.
thanks alot

Similar Messages

  • Create Workspace without demo-application and tables

    Is it possible to crate a new workspace on a existing schema without creating new tables and demo-applications?
    Thanks
    Mario

    Mario,
    This is currently not possible.
    Sergio

  • Slow split table export (R3load and WHERE clause)

    For our split table exports, we used custom coded WHERE clauses. (Basically adding additional columns to the R3ta default column to take advantage of existing indexes).
    The results have been good so far. Full tablescans have been eliminated and export times have gone down, in some cases, tables export times have improved by 50%.
    However, our biggest table, CE1OC01 (120 GB), continues to be a bottleneck. Initially, after using the new WHERE clause, it looked like performance gains were dramatic, with export times for the first 5 packages dropping from 25-30 hours down to 1 1/2 hours.
    However, after 2 hours, the remaining CE1OC01 split packages have shown no improvement. This is very odd because we are trying to determine why part of the table exports very fast, but other parts are running very slow.
    Before the custom WHERE clauses, the export server had run into issues with SORTHEAP being exhausted, so we thought that might be the culprit. But that does not seem to be an issue now, since the improved WHERE clauses have reduced or eliminated excessive sorting.
    I checked the access path of all the CE1OC01 packages, through EXPLAIN, and they all access the same index to return results. The execution time in EXPLAIN returns similar times for each of the packages:
    CE1OC01-11: select * from CE1OC01  WHERE MANDT='212'
    AND ("BELNR" > '0124727994') AND ("BELNR" <= '0131810250')
    CE1OC01-19: select * from CE1OC01 WHERE MANDT='212'
    AND ("BELNR" > '0181387534') AND ("BELNR" <= '0188469413')
          0 SELECT STATEMENT ( Estimated Costs =  8.448E+06 [timerons] )
      |
      ---      1 RETURN
          |
          ---      2 FETCH CE1OC01
              |
              ------   3 IXSCAN CE1OC01~4 #key columns:  2
    query execution time [millisec]            |       333
    uow elapsed time [microsec]                |   429,907
    total user CPU time [microsec]             |         0
    total system cpu time [microsec]           |         0
    Both queries utilize an index that has fields MANDT and BELNR. However, during R3load, CE1OC01-19 finishes in an hour and a half, whereas CE1OC01-11 can take 25-30 hours.
    I am wondering if there is anything else to check on the DB2 access path side of things or if I need to start digging deeper into other aggregate load/infrastructure issues. Other tables don't seem to exhibit this behavior. There is some discrepancy between other tables' run times (for example, 2-4 hours), but those are not as dramatic as this particular table.
    Another idea to test is to try and export only 5 parts of the table at a time, perhaps there is a throughput or logical limitation when all 20 of the exports are running at the same time. Or create a single column index on BELNR (default R3ta column) and see if that shows any improvement.
    Anyone have any ideas on why some of the table moves fast but the rest of it moves slow?
    We also notice that the "fast" parts of the table are at the very end of the table. We are wondering if perhaps the index is less fragmented in that range, a REORG or recreation of the index may do this table some good. We were hoping to squeeze as many improvements out of our export process as possible before running a full REORG on the database. This particular index (there are 5 indexes on this table) has a Cluster Ratio of 54%, so, perhaps for purposes of the export, it may make sense to REORG the table and cluster it around this particular index. By contrast, the primary key index has a Cluster Ratio of 86%.
    Here is the output from our current run. The "slow" parts of the table have not completed, but they average a throughput of 0.18 MB/min, versus the "fast" parts, which average 5 MB/min, a pretty dramatic difference.
    package     time      start date        end date          size MB  MB/min
    CE1OC01-16  10:20:37  2008-11-25 20:47  2008-11-26 07:08   417.62    0.67
    CE1OC01-18   1:26:58  2008-11-25 20:47  2008-11-25 22:14   429.41    4.94
    CE1OC01-17   1:26:04  2008-11-25 20:47  2008-11-25 22:13   416.38    4.84
    CE1OC01-19   1:24:46  2008-11-25 20:47  2008-11-25 22:12   437.98    5.17
    CE1OC01-20   1:20:51  2008-11-25 20:48  2008-11-25 22:09   435.87    5.39
    CE1OC01-1    0:00:00  2008-11-25 20:48                       0.00
    CE1OC01-10   0:00:00  2008-11-25 20:48                     152.25
    CE1OC01-11   0:00:00  2008-11-25 20:48                     143.55
    CE1OC01-12   0:00:00  2008-11-25 20:48                     145.11
    CE1OC01-13   0:00:00  2008-11-25 20:48                     146.92
    CE1OC01-14   0:00:00  2008-11-25 20:48                     140.00
    CE1OC01-15   0:00:00  2008-11-25 20:48                     145.52
    CE1OC01-2    0:00:00  2008-11-25 20:48                     184.33
    CE1OC01-3    0:00:00  2008-11-25 20:48                     183.34
    CE1OC01-4    0:00:00  2008-11-25 20:48                     158.62
    CE1OC01-5    0:00:00  2008-11-25 20:48                     157.09
    CE1OC01-6    0:00:00  2008-11-25 20:48                     150.41
    CE1OC01-7    0:00:00  2008-11-25 20:48                     175.29
    CE1OC01-8    0:00:00  2008-11-25 20:48                     150.55
    CE1OC01-9    0:00:00  2008-11-25 20:48                     154.84

    Hi all, thanks for the quick and extremely helpful answers.
    Beck,
    Thanks for the health check. We are exporting the entire table in parallel, so all the exports begin at the same time. Regarding the SORTHEAP, we initially thought that might be our problem, because we were running out of SORTHEAP on the source database server. Looks like for this run, and the previous run, SORTHEAP has remained available and has not overrun. That's what was so confusing, because this looked like a buffer overrun.
    Ralph,
    The WHERE technique you provided worked perfectly. Our export times have improved dramatically by switching to the forced full tablescan. Being always trained to eliminate full tablescans, it seems counterintuitive at first, but, given the nature of the export query, combined with the unsorted export, it now makes total sense why the tablescan works so much better.
    Looks like you were right, in this case, the index adds too much additional overhead, and especially since our Cluster Ratio was terrible (in the 50% range), so the index was definitely working against us, by bouncing all over the place to pull the data out.
    We're going to look at some of our other long running tables and see if this technique improves runtimes on them as well.
    Thanks so much, that helped us out tremendously. We will verify the data from source to target matches up 1 for 1 by running a consistency check.
    Look at the throughput difference between the previous run and the current run:
    package     time       start date        end date          size MB  MB/min
    CE1OC01-11   40:14:47  2008-11-20 19:43  2008-11-22 11:58   437.27    0.18
    CE1OC01-14   39:59:51  2008-11-20 19:43  2008-11-22 11:43   427.60    0.18
    CE1OC01-12   39:58:37  2008-11-20 19:43  2008-11-22 11:42   430.66    0.18
    CE1OC01-13   39:51:27  2008-11-20 19:43  2008-11-22 11:35   421.09    0.18
    CE1OC01-15   39:49:50  2008-11-20 19:43  2008-11-22 11:33   426.54    0.18
    CE1OC01-10   39:33:57  2008-11-20 19:43  2008-11-22 11:17   429.44    0.18
    CE1OC01-8    39:27:58  2008-11-20 19:43  2008-11-22 11:11   417.62    0.18
    CE1OC01-6    39:02:18  2008-11-20 19:43  2008-11-22 10:45   416.35    0.18
    CE1OC01-5    38:53:09  2008-11-20 19:43  2008-11-22 10:36   413.29    0.18
    CE1OC01-4    38:52:34  2008-11-20 19:43  2008-11-22 10:36   424.06    0.18
    CE1OC01-9    38:48:09  2008-11-20 19:43  2008-11-22 10:31   416.89    0.18
    CE1OC01-3    38:21:51  2008-11-20 19:43  2008-11-22 10:05   428.16    0.19
    CE1OC01-2    36:02:27  2008-11-20 19:43  2008-11-22 07:46   409.05    0.19
    CE1OC01-7    33:35:42  2008-11-20 19:43  2008-11-22 05:19   414.24    0.21
    CE1OC01-16    9:33:14  2008-11-20 19:43  2008-11-21 05:16   417.62    0.73
    CE1OC01-17    1:20:01  2008-11-20 19:43  2008-11-20 21:03   416.38    5.20
    CE1OC01-18    1:19:29  2008-11-20 19:43  2008-11-20 21:03   429.41    5.40
    CE1OC01-19    1:16:13  2008-11-20 19:44  2008-11-20 21:00   437.98    5.75
    CE1OC01-20    1:14:06  2008-11-20 19:49  2008-11-20 21:03   435.87    5.88
    PLPO          0:52:14  2008-11-20 19:43  2008-11-20 20:35    92.70    1.77
    BCST_SR       0:05:12  2008-11-20 19:43  2008-11-20 19:48    29.39    5.65
    CE1OC01-1     0:00:00  2008-11-20 19:43                       0.00
                558:13:06  2008-11-20 19:43  2008-11-22 11:58  8171.62
    package     time      start date        end date          size MB   MB/min
    CE1OC01-9    9:11:58  2008-12-01 20:14  2008-12-02 05:26   1172.12    2.12
    CE1OC01-5    9:11:48  2008-12-01 20:14  2008-12-02 05:25   1174.64    2.13
    CE1OC01-4    9:11:32  2008-12-01 20:14  2008-12-02 05:25   1174.51    2.13
    CE1OC01-8    9:09:24  2008-12-01 20:14  2008-12-02 05:23   1172.49    2.13
    CE1OC01-1    9:05:55  2008-12-01 20:14  2008-12-02 05:20   1188.43    2.18
    CE1OC01-2    9:00:47  2008-12-01 20:14  2008-12-02 05:14   1184.52    2.19
    CE1OC01-7    8:54:06  2008-12-01 20:14  2008-12-02 05:08   1173.23    2.20
    CE1OC01-3    8:52:22  2008-12-01 20:14  2008-12-02 05:06   1179.91    2.22
    CE1OC01-10   8:45:09  2008-12-01 20:14  2008-12-02 04:59   1171.90    2.23
    CE1OC01-6    8:28:10  2008-12-01 20:14  2008-12-02 04:42   1172.46    2.31
    PLPO         0:25:16  2008-12-01 20:14  2008-12-01 20:39     92.70    3.67
                90:16:27  2008-12-01 20:14  2008-12-02 05:26  11856.91

  • Doubt in  export, import and table para when creating Function Module

    Dear fellow ABAPers,
    I have a doubt in defining export, import and table parameter while creating a function module.
    I am calling a function module inside a user exit. Now in the user exit the SAP fills an internal table called i_lfa1 with all the data user has eneterd.
    Now I want to pass this whole internal table to function module and the perform some checks on the values of internal table.
    After that function module fills an error structure with values depending on some check.
    1)
    How do I pass this internal table to function module ? 
    When I am creating function module in se37 where do I define this iternal table type ? Is it in Import or Table parameter during function module creation?
    2)
    Where do I define error structure type (which is returned by function module to main program)? Is it in Export or table parameter during function module creation?
    Please clear my doubt..
    Relevant points will be awarded.
    Regards,
    Tushar.

    Hi Tushar,
    1. How do I pass this internal table to function module ?
       I assume u are creating your own Y/Z FM.
       Pass it thru TABLES parameter.
    2. When I am creating function module in se37 where do I define this iternal table type
       Define this in TABLES interface.
       What Type ?
       THE SAME TYPE WHICH HAS BEEN DEFINED
        WHILE PASSING IN THE USER-EXIT FUNCTION MODULE.
       IF U SEE THE FM OF THE USER-EXIT,
       U WILL COME TO KNOW.
    3.
    Where do I define error structure type (which is returned by function module to main program)? Is it in Export or table parameter during function module creation?
    Define it in TABLES interace. (not in export, import)
      (Since what u are going to return is an internal table)
      U can take for eg. BDCMSGCOLL.
      OR u can create your own Y/Z structure
    for the same purpose.
      (or u can use the structure type T100)
    I hope it helps.
    Regards,
    Amit M.

  • Setting up Figure and Table Titles for Translation Import and Export?

    Hi,
    I use InDesign to write User Manuals. Once the manual complete, I use SDL World Server to translate the document which works wonderful. The way it works is the document is parsed by the xml data and when you export the document from SDL, the translated data replaces the English data, it’s pretty slick! The problem that I have is that my Figure and Table numbers are hard coded with the word "Figure" and "Table” as part of the numbering system and will not translate nor export (see screenshot below). Is there A.) A way I can make the code become a reference to a text box in the document that then can be translated? B.) A common or different way to handle this?
    If you need any more info, just let me know. Thank you for your help!

    The problem that I have is that my Figure and Table numbers are hard coded with the word "Figure" and "Table” as part of the numbering system and will not translate nor export (see screenshot below). Is there A.) A way I can make the code become a reference to a text box in the document that then can be translated? B.) A common or different way to handle this?
    I've always avoided that auto-figure-naming stuff, in any application, for the sole reason that it often doesn't localize well. I'm not a WorldServer user, but a SDL Trados Studio user, so I don't know exactly what to tell you. If you're the kind of person to pop your IDML open in a text editor, you'll find that the actual "Figure" text appears in Styles.xml, and I doubt that SDL's IDML filters look there for text for translation. I don't actually know for sure, and I do know you're using an SDL product with which I have no familiarity whatsoever.
    A different, somewhat kludgy way to handle this would be to use auto-numbering to create an invisible paragraph number (color of Paper, size of 0.1 pts, glyph width of 1%, or something along those lines), and then have the phrase "Figure" in your live text, followed by a cross-reference to the paragraph number.
    A common way to handle this would be to ask SDL to update their IDML filter to pick up terms used in auto-numbered lists.

  • Oracle11gR2 Workspace Manager and table consistency after merge

    Hi folks,
         I'm working with Oracle Workspace Manager in order to get data inserted and validated into workpaces before they become available to the LIVE workspace.
         Doing some tests I found a problem about data consistency after I merge the data from a child workspace to the parent workspace.
         To be able to explain and reproduce the problem I create a simple test case:
    --Create table TB_LINK
    create table TB_LINK
      CD_LINK NUMBER not null,
      DS_LINK VARCHAR2(30)
    --Create primary key
    alter table TB_LINK add primary key (CD_LINK);
    --Create table TB_GUD
    create table TB_GUD
      CD_GUD  NUMBER not null,
      DS_GUD  VARCHAR2(30),
      CD_LINK NUMBER
    -- Create primary key
    alter table TB_GUD add primary key (CD_GUD);
    -- Create foreign key 
    alter table TB_GUD
      add constraint FK_TB_LINK foreign key (CD_LINK)
      references TB_LINK (CD_LINK);
    -- Create sequences
    create sequence SEQ_TB_GUD
    minvalue 1
    maxvalue 9999999999999999999999999999
    start with 1
    increment by 1
    nocache;
    create sequence SEQ_TB_LINK
    minvalue 1
    maxvalue 9999999999999999999999999999
    start with 1
    increment by 1
    nocache; 
    --Create Triggers
    create or replace trigger "INS_TB_GUD" before insert on TB_GUD for each row
    Begin
    select SEQ_TB_GUD.nextval into :new.CD_GUD from dual;
    end;
    create or replace trigger "INS_TB_LINK" before insert on TB_LINK for each row
    Begin
    select SEQ_TB_LINK.nextval into :new.CD_LINK from dual;
    end;
    --Enable version TB_LINK and TB_GUD
    EXECUTE DBMS_WM.EnableVersioning('TB_GUD','VIEW_WO_OVERWRITE',FALSE,FALSE,'UNLIMITED');
    EXECUTE DBMS_WM.EnableVersioning('TB_LINK','VIEW_WO_OVERWRITE',FALSE,FALSE,'UNLIMITED');
    --Create a workspace
    EXECUTE DBMS_WM.CreateWorkspace ('TEST_WKS');
    --Goto workspace TEST_WKS
    EXECUTE dbms_wm.gotoworkspace('TEST_WKS');
    --Insert data into TB_LINK and TB_GUD
    INSERT INTO TB_LINK(DS_LINK) VALUES ('DS1');
    INSERT INTO TB_LINK(DS_LINK) VALUES ('DS2');
    INSERT INTO TB_LINK(DS_LINK) VALUES ('DS3');
    INSERT INTO TB_LINK(DS_LINK) VALUES ('DS4');
    COMMIT;
    INSERT INTO TB_GUD(DS_GUD,CD_LINK) VALUES ('GUD1',1);
    INSERT INTO TB_GUD(DS_GUD,CD_LINK) VALUES ('GUD2',2);
    INSERT INTO TB_GUD(DS_GUD,CD_LINK) VALUES ('GUD3',3);
    INSERT INTO TB_GUD(DS_GUD,CD_LINK) VALUES ('GUD4',4);
    COMMIT;
    --Checking keys
    select * from tb_link;
       CD_LINK      DS_LINK
             1           DS1
             2           DS2
             3           DS3
             4           DS4
    select * from tb_gud;
       CD_GUD      DS_GUD     CD_LINK
             1           GUD1              1
             2           GUD2              2
             3           GUD3              3
             4           GUD4              4
    --Merge Workspace
    EXECUTE DBMS_WM.MergeWorkspace ('TEST_WKS');
    --Checking keys
    EXECUTE dbms_wm.gotoworkspace('LIVE');
    select * from tb_link;
       CD_LINK      DS_LINK
             5           DS4
             6           DS3
             7           DS1
             8           DS2
    We can see that the CD_LINK got new values after merge and that was not expected.
    select * from tb_gud;
       CD_GUD      DS_GUD                  CD_LINK
             6           GUD3                       3
             7           GUD1                       1
             8           GUD2                       2
             5           GUD4                       4
    We can see that the CD_GUD got new values after merge and that was not expected.
    Now, the values for the CD_LINK column does not have corresponding records at the TB_LINK table, as the foreign key does not exist anymore.
    Could you please help me understand what is going on?
    Thanks,
    Luis

    Hi Luis,
    The reason for the difference is that the trigger is being run during the MergeWorkspace operation.  The inserts into the child workspace(TEST_WKS) translates into an insert into the LIVE workspace during merge as the rows do not yet exist.  As a result, the trigger is fired and the sequence is evaluated.  Ideally, we should not allow the PK to be modified by a sequence in this case.
    You have 2 options:
    (1) Check for :new.CD_GUD being null prior to using the sequence.  Any dml coming from a merge/refresh operation will have a non-null value.
    (2) Turn off the trigger during dbms_wm procedures.  This can be done using dbms_wm.SetTriggerEvents.  I would assume you would only want this trigger being run for DML events.
    Let me know if you have any questions.
    Regards,
    Ben

  • My hard drive crashed on my Mac Book Pro and I just had it replaced yesterday.  After restoring my data, Safari will not open.  The error message I continue to receive is "Safari quit unexpectedly"  Should I just delete the application and download a new?

    My hard drive crashed on my Mac Book Pro and I just had it replaced yesterday.  After restoring my data, Safari will not open.  The error message I continue to receive is "Safari quit unexpectedly"  Should I just delete the application and download a new one?

    http://swcdn.apple.com/content/downloads/37/23/041-5058/lr5tynbldi18zcrqo8a8uq88 rnjushqliu/Safari6Lion.pkg 
    Here you go.....

  • Ios7 bookamrking web application to home screen via Safari crashes after login and when we try to open any link inside. Can anyone please tell me if there is any scheduled fix for this bug in safari?

    ios7 bookmarking web application to home screen via Safari, crashes after login and when we try to open any link inside from the icon that gets created on the home screen. Can anyone please tell me if there is any scheduled fix for this bug in safari?
    From what i understand i see that - Cookies are not transferred between your website and your webapp when installing the icon on the home screen (for authentication purposes for example). It was working until 6.1 and now it’s not working anymore.
    Can someone please tell me if this bug will be fixed in any future release, if so the ticket number of some kind that i can track, or if there is a workaround for this issue?

    ios7 bookmarking web application to home screen via Safari, crashes after login and when we try to open any link inside from the icon that gets created on the home screen. Can anyone please tell me if there is any scheduled fix for this bug in safari?
    From what i understand i see that - Cookies are not transferred between your website and your webapp when installing the icon on the home screen (for authentication purposes for example). It was working until 6.1 and now it’s not working anymore.
    Can someone please tell me if this bug will be fixed in any future release, if so the ticket number of some kind that i can track, or if there is a workaround for this issue?

  • Error executing the Query after deploying and executing the application in WLS 6.1 SP3

    Hi,
    We are trying to run a application by deploying a war file on Web Logic Server
    6.1 SP3. After deploying and at the time of execution of the application, the
    server console displays this error message
    java.sql.SQLException: ORA-00600: internal error code, arguments: [ttcgcshnd-1],
    [0], [], [], [], [], [], []
    While debugging thru the application we found that this error comes at the time
    of execution of the query thru the Statement.executeQuery method of java.sql.Statement
    class.
    This same application when deployed on Web Logic Server 8.1 runs perfectly, without
    any hitches.
    Please, provide me the solution to this asap.

    Look for sometime around November. No exact date.
    Eric
    "Dominic Tulley" <[email protected]> wrote in message
    news:[email protected]..
    Thanks Eric,
    you know what I'm going to ask next right?
    Any ideas when SP4 comes out?
    Cheers,
    -Dominic
    "Eric Gross" <[email protected]> wrote in message
    news:[email protected]..
    The next version of Apache that we will support will be 2.0.42/2.0.43
    and
    the module for that will be included in the next Service Pack for6.1(SP4)
    and 7.0(SP2).
    The problem with Apache 2 is that when a new release comes out a newmodule
    needs to be compiled. Most of the time. For 2.0.42/2.0.43 this is notthe
    case.
    Regards,
    Eric
    "Dominic Tulley" <[email protected]> wrote in
    message
    news:3da3f246$[email protected]..
    I'm trying to set this up (initially just apache in front of a single
    WLS
    server but ultimately I want to put it in front of a cluster).
    I've installed Apache 2.0.40.
    I've copied the mod_wl_20.so file into the apache modules folder.
    I've edited the httpd.conf file and added the line:
    LoadModule weblogic_module modules/mod_wl_20.so
    When I run apache -t to check the configuration I get the following:
    C:\Program Files\Apache Group\Apache2\bin>apache -t
    Syntax error on line 173 of C:/Program Files/Apache
    Group/Apache2/conf/httpd.conf:
    Cannot load C:/Program Files/Apache Group/Apache2/modules/mod_wl_20.sointo
    server: The specified procedure could not be found.
    So what's going on here?
    Looking at previous postings it sounds like there's an issue with
    versions
    of apache after 2.0.39 but I can't find a download for that version.Also,
    it sounded like the issues were for WLS 7, not 6.1. Can this work
    with
    2.0.40 or am I wasting my time?
    If I get the mod_wl_20.so from dev2dev I get the "incompatible plugin"
    message that has been mentioned in this group already.
    Any suggestions appreciated,
    -Dominic

  • I installed the system update to 10.7.3. After re-booting, my mail file was missing from the dock.  Looked in applications and the mail application is missing. Any thoughts?

    I installed the system update to 10.7.3. After re-booting, my mail file was missing from the dock.  Looked in applications and the mail application is missing. Any thoughts?

    Thanks Paul; but, taking your suggestion didn't solve the problem.  As mail is part of the OS, the updates don't have the capacity to reinstall applications; so, I had to reinstall the OS.  I now have my mail back (although it didn't appear on the dock--I had to go to applications and activate it and then put it back in the dock.  Thanks again for trying.

  • CS5.5 and OSX 10.9.4: Most applications will not launch even after uninstalling and reinstalling the Creative Suite 5.5. After my 2011 MacBook Pro 17 died, I had to purchase a new MacBook Pro 15 to replace it.  I performed a Time Machine installation from

    CS5.5 and OSX 10.9.4: Most applications will not launch even after uninstalling and reinstalling the Creative Suite 5.5. After my 2011 MacBook Pro 17 died, I had to purchase a new MacBook Pro 15 to replace it.  I performed a Time Machine installation from that 2011 machine, and only Distiller and Acrobat would launch (none of the rest of the creation apps).  So I ran the CS5.5 installer, uninstalled, then reinstalled, producing the same result. When I try to launch, nothing happens.  What next?

    Try...
    Adobe Creative Suite Cleaner Tool
    helps resolve installation problems for CS3 thru CS6 and for Creative Cloud
    http://www.adobe.com/support/contact/cscleanertool.html

  • Am running Tiger 10.4.11; just got Snow Leopard 10.6.3 (regular OS disk, not an upgrade); when I install Snow Leopard, will it wipe everything?  overwrite Tiger?  Do I need to make backups of applications and files to reinstall after Sno Lprd loads? thx!

    Am running  Tiger 10.4.11; just got Snow Leopard 10.6.3 (regular OS disk, not upgrade).  When I install Snow Leopard, will it wipe everything?  overwrite Tiger?  Do I need to make backups of applications and files to reinstall after Snow Leopard Loads?  thanks!

    Hi Donald,
    No it doesn't erase anything by default, it does what we used to call an  Archive and Install feature, which can give you a new OS, yet preserve your files and settings if you have enough room on your HD...
    http://docs.info.apple.com/article.html?artnum=107120
    But it does it automatically unless you specifically erase the Disk, but will replace Tiger completely for the OS & several Apple Apps like Safari, Mail, etc, yet preserve your old settings for those!
    If you think you may need Tiger to run any APPs that do not work in SL/10.6, then clone or backup first.
    I would recomend 2 things before installing though...
    Make sure you have enough RAM & HD space.
    Snow Leopard/10.6.x Requirements...
    General requirements
       * Mac computer with an Intel processor
        * 1GB of memory (I say 2GB at least)
        * 5GB of available disk space (I say 30GB at least)
    If it's a core Duo & not a Core2Duo, then it'll only run in 32 bit mode.
    "Try Disk Utility
    1. Insert the new Mac OS X SL/10.6 Install Disc, then restart the computer while holding the C key.
    2. When your computer finishes starting up from the disc, choose Disk Utility from the Installer menu. (In Mac OS X 10.4 or later, you must select your language first.)
    Important: Do not click Continue in the first screen of the Installer. If you do, you must restart from the disc again to access Disk Utility.
    3. Click the First Aid tab.
    4. Select your Mac OS X volume.
    5. Click Repair. Disk Utility checks and repairs the disk."
    http://docs.info.apple.com/article.html?artnum=106214

  • There is an inconsistency between the authentication mode of target web application and the source web application after migrating to claims

    I've had my farm upgraded from SP2010 to SP2013 for over 6 months now and all is well, however, I was refreshing my staging environment from production and I noticed that one of the databases still shows these errors when I run test-spcontentdatabase:
    Category             : Configuration
    Error             : False
    UpgradeBlocking : False
    Message           : The [SharePoint Web App] web application is configured with claims authentication mode however the content database you are trying to attach is intended to be used against
    a windows classic authentication mode.
    Remedy              : There is an inconsistency between the authentication mode of target web application and the source web application. Ensure that the authentication mode setting in upgraded web application is the
    same as what you had in previous SharePoint 2010 web application. Refer to the link "http://go.microsoft.com/fwlink/?LinkId=236865" for more information.
    This doesn't make sense considering I converted the production web application to claims during the upgrade and then verified all sites were working with claims logins. I also verified that existing AD user identities were converted to claims by checking out
    the database tables. Yet test-spcontentdatabase still thinks there is a mismatch here.
    My farm is SP1 and no further CUs. The point of this particular refresh is so I can update to the November CUs in my test farm. Anyone else see this? Seems like it's a bug/safe to ignore because my stuff is working.
    Thanks,
    Aaron

    See:
    http://thesharepointfarm.com/2014/11/test-spcontentdatabase-classic-to-claims-conversion/
    Trevor Seward
    Follow or contact me at...
    &nbsp&nbsp
    This post is my own opinion and does not necessarily reflect the opinion or view of Microsoft, its employees, or other MVPs.

  • I recently upgraded from Leopard to Snow Leopard and I noticed that I have more capacity available on the Macintosh HD. I checked my files and applications and they are still on my computer. Can capacity on the HD change after an OS upgrade?

    I recently upgraded from Leopard to Snow Leopard on my MacBook and I noticed that I have more capacity on the Macintosh HD than before the upgrade. I checked my files and applications and they are still on my computer and nothing appears to be missing. Before the upgrade I had around 111 GB capacity with around 10.50 GB available on the Mac HD but now after the upgrade I have 119 GB capacity and 20.92 GB available. Can capacity on the HD change after an OS upgrade?

    No, but the upgrade removed unneeded PowerPC code and changed the way free space is determined.
    (72071)

  • I'm trying to install a cd application and I keep getting the message 'you do not have permission to use the application FSP OS X'  Even after selecting always allow and entering my login details it still won't accept it.  Help!!..

    I'm trying to install a cd application and I keep getting the message 'you do not have permission to use the application FSP OS X'  Even after selecting always allow and entering my login details it still won't accept it.  Help!!..

    Please repost in the Final Cut Pro forum. Have you considered upgrading your machine, if you are still running Leopard your machine may be as old as 5 years old. While this should not matter you probably won't get the performance you want out of a demanding app like FCP.
    Best of luck.

Maybe you are looking for

  • Iweb not opening at all

    The application Iweb quit unexpectedly, Mac OS X and other applicatioms are not affected. Click Relaunch to launch the application again. I was working on a new website and downloaded and was testing out a bunch of new fonts, when i tried to open Iwe

  • Acrobat Standard 9.1.1 crashes on Create PDF from scanner

    All options under the Create PDF from scanner sends Acrobat Standard 9.1.1 crashing. Faulting application Acrobat.exe, version 9.1.0.163, time stamp 0x49a88b32, faulting module dlbatwds.ds, version 3.107.0.0, time stamp 0x45df4d3d, exception code 0xc

  • This may sound ridiculous but how do you stop a song playing ?

    I can play a song, pause a song, shuffle and many other things but I can not find any function that just stops it altogether! In video you press the "done" button but no such thing appears in music - am I going mad?!

  • OIM 11g: Custom Scheduler code

    Hi I need to develop custom scheduler in OIM 11g. But I am struck at how to use execute() method. public class CustomScheduler extends oracle.iam.scheduler.vo.TaskSupport { *//How to use execute(HashMap argMap) -- What hashmap data I need to pass to

  • OPM Actual Cost Process not selecting all Inventory Movements

    We have completed 27 Move Immediate transactions (TRNI) but only 2 of the transactions appear in the Actual Cost Transactions View so our PMAC cost is not as expected. Those 2 transactions were in Sept '09. All other transactions happened in Apr, May