Best way to handle raiseerror after migration from SQL 2K?

Hi,
Here a sample code :
raise_application_error(-20999, 'message id: 50048');
/*[SPCONV-ERR(7)]:('Raiserror ( <message id>....) Manual conversion required*/
/* message id gives id in sqlserver syscomments table*/
RAISE_APPLICATION_ERROR(-20500,'Trigger forces a rollback');
/* ROLLBACK; */
I wonder what the best way to find the equivalent message id, or what the best to do in oracle to replace those event....???
Thanks in advance
Pete

No matter found something !

Similar Messages

  • Best way to handle duplicate headings stemming from linked TOC book?

    What's the best way to handle duplicate topic titles stemming from TOC books that contain links to a topic that you want to have appear in the body? The problem I've had for years now is that the TOC generates one heading, and the topic itself generates one heading.This results in duplicate headings in the printed output.
    I have a large ~2500 topic project I have to print every release, and to date we've been handling it in post-build word macros, but it's not 100% effective, so we're looking to fix this issue on the source side of the fence. On some of our smaller projects, we've actually marked with the heading in the topic itself with the Online CBT and that seems to work. We're thinking of doing the same to our huge project unless there's a better way of handling this. Any suggestions?

    See the tip immediately above this link. http://www.grainge.org/pages/authoring/printing/rh9_printing.htm#wizard_page3
    The alternative is to remove the topic from the print layout so that it only generates by virtue of its link to the book.
    See www.grainge.org for RoboHelp and Authoring tips
    @petergrainge

  • Best way to refresh page after returning from task flow?

    Hello -
    (Using jdev 11g release 1)
    What is the best way to refresh data in a page after navigating to and returning from a task flow with an isolated data control scope where that data is changed and commited to the database?
    I have 2 bounded task flows: list-records-tf and edit-record-tf
    Both use page fragments
    list-records-tf has a list.jsff fragment and a task flow call to edit-record-tf
    The list.jsff page has a table of records that a user can click on and a button which, when pressed, will pass control to the edit-record-tf call. (There are also set property listeners on the button to set values in the request that are used as parameters to edit-record-tf.)
    The edit-record-tf always begins a new transaction and does not share data controls with the calling task flow. It consists of an application module call to set up the model according to the parameters passed in (edit record X or create new record Y or...etc.), a page fragment with a form to allow users to edit the record, and 2 different task flow returns for saving/cancelling the transaction.
    Back to the question - when I change a record in the edit page, the changes do not show up on the list page until I requery the data set. What is the best way to get the list page to refresh itself automatically upon return from the edit-record-tf?
    (If I ran the edit task flow in a popup dialog I could just use the return listener on the command component that launched the popup. But I don't want to run this in a dialog.)
    Thank you for reading my question.

    What if you have the bean which has refresh method as TF param? Call that method after you save the data. or use contextual event.

  • What is the best way to handle executing multiple packages from the Agent?

    I have several packages that have to be executed in sequence. I thought the best way to do that was by creating a job for each package then have a master job that executes the other packages. In my master job,  I'm using sp_start_job to call the other
    jobs. The problem is, the master job moves from step to step without waiting for the child jobs to finish; basically they all execute together.
    That is the way I've seen it done in other places so I feel like I'm doing something wrong. In the alternative, I know it's possible to set the individual steps up so they execute the packages directly without calling an external job. I prefer the first
    way though.
    Which way should I jump on this?

    So basically what I'm hearing is just call the packages in a mulit step job. Creating a master package and calling child packages sounds a little crazy and unscaleable especially considering that the packages have master child relationships
    within themselves. It's SSIS Package Inception. 
    Sorry whats the issue with that?
    Provided you're setting the package sequence correctly based on your dependency it will work fine as loop gets iterated based on how you set the package list.
    What we have is a audit and control mechanism which even have details on the dependency so based dependency set only the packages gets ordered and listed for the loop to iterate through and executing. Also tomorrow if a new package comes into being, all
    it takes for us is to tweak the audit table to add a new entry for the package and set the dependency and it will continue to work fine including the new package without touching the existing job for any modification whatsoever.
    Another advantage of this table is we also capture audit details in it like date when package got last executed, its status of execution,rows processed (insert,modified, deleted) etc which can be used for easy monitoring of the data processing tasks as well.
    Please Mark This As Answer if it solved your issue
    Please Vote This As Helpful if it helps to solve your issue
    Visakh
    My Wiki User Page
    My MSDN Page
    My Personal Blog
    My Facebook Page

  • Replication Configuration issue after migrating from SQL Server 2005 to SQL Server 2008 R2

    Hi All,
    This weekend we have migrated one of our Core system from SQL Server 2005 SP3 to SQL Server 2008 R2 SP2. Migration went successful. However when we are trying to Configure transaction replication we are getting below Error:
    Msg 21854, Level 16, State 1, Procedure sp_MSrepl_addarticle, Line 573
    Could not add new article to publication 'abc' because of active schema change activities or a snapshot is being generated.
    Background:
    Database compability Changed from 90 to 100.
    A quick Help would be appreciated.
    Best Regards - Viral
    Thanks - Viral

    Is a snapshot being generated? the message seems to indicate one is.
    looking for a book on SQL Server 2008 Administration?
    http://www.amazon.com/Microsoft-Server-2008-Management-Administration/dp/067233044X looking for a book on SQL Server 2008 Full-Text Search?
    http://www.amazon.com/Pro-Full-Text-Search-Server-2008/dp/1430215941

  • SQL 2005 x64 poor performance after migration from SQL 2005 x86

    Recently We've migrated SQL 2005 x86 SP1 Enterprise running on Windows 2003 Enterprise to new hardware where is running Windows 2008 SP2 and SQL 2005 x64 SP4 with CU3.
    In the work day, there are sometime SQL excpetions like this below. On other hand i can also tell that we already found some queries which has some inner joins and where clauses on nvarchar fileds, because of the nature of application. Such a query
    is running for about 11 seconds on old SQL server, and about 6 minutes on new SQL server. My main question if is this possible just with a change to new architecture of SQL and more last patches and updates for SQL 2005. Is there any way how to get
    this queries running faster, without rewritng them with new ones? Is this also possible this could be also cause for SQL exceptions, which we are facing on SQL server?
    About new hardware, for now we can't find it something wrong. There is basic configuration:
    Cisco UCS  C220 M3S (1U server)
    32GB RAM
    Intel(R)Xeon® CPU [email protected], 6 core
    Disk system: 4xSAS 300GB in RAID5 (OS and DATA) and 2xSATA 1TB in RAID1 (SQL backup)
    RAID controller: LSI MegaRAID SAS 9271-8i (with latest driver and firmware)
    Part of log of dump file:
    2014-05-09 08:30:07.08 spid86      Error: 17311, Severity: 16, State: 1.
    2014-05-09 08:30:07.08 spid86      SQL Server is terminating because of fatal exception c0000264. This error may be caused by an unhandled Win32 or C++ exception, or by an access violation encountered during exception handling.
    Check the SQL error log for any related stack dumps or messages. This exception forces SQL Server to shutdown. To recover from this error, restart the server (unless SQLAgent is configured to auto restart).
    2014-05-09 08:30:07.08 spid86      Using 'dbghelp.dll' version '4.0.5'
    2014-05-09 08:30:07.08 spid86      **Dump thread - spid = 86, PSS = 0x0000000132A67ED0, EC = 0x0000000132A67EE0
    2014-05-09 08:30:07.08 spid86      ***Stack Dump being sent to D:\MSSQL.1\MSSQL\LOG\SQLDump0087.txt
    2014-05-09 08:30:07.08 spid86      * *******************************************************************************
    2014-05-09 08:30:07.08 spid86      *
    Thanks for any help,
    Matjaž

    >Is there any way how to get this queries running faster, without rewriting them with new ones?
    You need to perform basic performance tuning steps in the new environment.
    Software optimization:
    http://www.sqlusa.com/articles/query-optimization/
    Disk optimization:
    http://www.sqlusa.com/bestpractices/configureharddisk/
    Memory optimization:
    http://www.sqlusa.com/bestpractices/memory-configuration/
    REBUILD INDEXES, eliminate missing indexes & UPDATE STATISTICS.
    If there is still a problem, look in Activity Monitor for expensive queries. Optimize them. As a general rule, you don't want to have expensive queries in RDBMS.
    Kalman Toth Database & OLAP Architect
    SQL Server 2014 Design & Programming
    New Book / Kindle: Exam 70-461 Bootcamp: Querying Microsoft SQL Server 2012

  • Implementing "Temporary Tables" in Oracle after migrating from SQL Server

    Hello,
    Some stored procedures referencing temporary tables in SQL Server were migrated to Oracle. In the Oracle DB, these stored procedures are invalid. In a bid to make them valid, I created the referenced temporary tables in my Oracle DB, via SQL*PLUS having logged into the database as the appropriate user. And when I query the user_tables view, I see the temporary tables that I created. Thereafter, when I compile the stored procedures, I still get the message: PL/SQL: ORA-00942: table or view does not exist. And the stored procedures remain invalid. Pls,what do I do ?
    Thanks for your cooperation

    Chux,
    The Migration Workbench should have created all the necessary temporary tables as part of the process to convert the stored procedures. Can you post the example part of the T/SQL stored procedure and what we converted it to?
    Our support for Temporary Tables is discussed in our reference guide.
    http://download-west.oracle.com/docs/html/B10254_01/ch5.htm#1026115
    Donal

  • Best way to handle an event from WaitForSingleObject

    Hello LV experts,
    I would like to know what's the best way to handle an event which I catch using the Win32 API (WaitForSingleObject) in Labview(7.1 or 8.20).
    I know there is too many possibilities to do this, but I want to do it  on the best way. that is why I would like to have ideas and tips from you.
    Any and all tips appreciated

    You can download the Library from this link. Inside is an example that shows how to handle windows messages/events.
    http://zone.ni.com/devzone/cda/epd/p/id/4394

  • (workflow question) - What is the best way to handle audio in a large Premiere project?

    Hey all,
    This might probably be suitable for any version of Premiere, but just in case, I use CS4 (Master Collection)
    I am wrestling in my brain about the best way to handle audio in my project to cut down on the time I am working on it.
    This project I just finished was a 10 minute video for a customer shot on miniDV (HVX-200) cut down from 3 hours of tape.
    I edited my whole project down to what looked good, and then I decided I needed to clean up all the Audio using Soundbooth, So I had to go in clip by clip, using the Edit in SoundBooth --> Render and Replace method on every clip. I couldn't find a way to batch edit any audio in Soundbooth.
    For every clip, I performed similar actions---
    1) both tracks of audio were recorded with 2 different microphones (2 mono tracks), so I needed only audio from 1 track - I used SB to cut and paste the good track over the other track.
    2) amplified the audio
    3) cleaned up the background noise with the noise filter
    I am sure there has to be a better workflow option than what I just did (going clip by clip), Can someone give me some advice on how best to handle audio in a situation like this?
    Should I have just rendered out new audio for the whole tape I was using, and then edit from that?
    Should I have rendered out the audio after I edited the clips into one long track and performed the actions I needed on it? or something entirely different? It was a very slow, tedious process.
    Thanks,
    Aza

    Hi, Aza.
    Given that my background is audio and I'm just coming into the brave new world of visual bits and bytes, I would second Hunt's recommendation regarding exporting the entire video's audio as one wav file, working on it, and then reimporting. I do this as one of the last stages, when I know I have the editing done, with an ear towards consistency from beginning to end.
    One of the benefits of this approach is that you can manage all audio in the same context. For example, if you want to normalize, compress or limit your audio, doing it a clip at a time will make it difficult for you to match levels consistently or find a compression setting that works smoothly across the board. It's likely that there will instead be subtle or obvious differences between each clip you worked on.
    When all your audio is in one file you can, for instance, look at the entire wave form, see that limiting to -6 db would trim off most of the unnecessary peaks, triim it down, and then normalize it all. You may still have to do some tweaking here and there, but it gets you much farther down the road, much more easily.Same goes for reverb, EQ or other effects where you want the same feel throughout the entire video.
    Hope this helps,
    Chris

  • Testing of SAP BW system after migration from OS Unix to windows

    Hi Experts,
    I am working on a project wherein testing of BW system is required after migration from OS Unix to Windows.
    If any one has done same set of activities in past or has any idea of this, please share all major test steps required after migration.
    Thanks in advance.
    Regards,
    Neeraj

    Hi Neeraj,
    I will paste the list we used for our upgrade from 3.1 to 7.0. I removed some company specific details. Sorry for the "lay-out", it is copied from excel, this messes up the whole message.
    best regards, André.
    Preparation     PHASE 1
    Review New Functionality     Review BW section on SAPNet for latest information on new release
    Confirm Upgrade Timing (First Sat)    
    GUI Upgrade     Install GUI's for Upgrade Test persons
    "Check Prerequisites     -Operating System
    -SAP Kernel
    -DBMS
    -Disk space
    -R/3 plugin
    -BW software version and SP
    Citrix.."
    "Check Compatibility Reqs with other systems     APO
    ECC6.0"
    Check Compatibility Reqs with 3rd party software     ?
    "Test Team 2004s Delta Training     To
    have idea on new functionality
    not have confusion if looks different.."
    "End User Communication: upgraded system and changes     e.g. right vs left click
    - send out communication twice"
    End User Communication: several "look and feel" sessions  NOTE: incorporate WISBECH !!!    
    Setup Portal connected to BW QA for testing once upgraded     Setup Portal connected to BW DEV for testing once upgraded
    Check best go live date - based on assuption 2 days needed    
    Determine test set  - based on input BS / KU    
    Estimate resources Business for testing    
    BluePrint     PHASE 2
    Check/CleanUp Development Objects     Check all open and not-released developments for release prior to upgrade
    Prepare System Setup     Prepare (if needed) flat-files for test loads
    Check OSS for release notes on objects     Each functional area should check out OSS for release notes on objects (for changes etcu2026, new functionality)
    "Prepare (detail) list of objects to be tested     Complete list of objects to be tested, special attention to custom/non-standard programs, enhancements. List of:
    - extractors
    - process chains
    - reports"
    "Setup delta mechanisms (create transaction data)     BackEnd Test: Financial Extraction and load, process chains
    u2022tbd
    u2022
    u2022 "
    Verification / creation of process chains     Check existence of to-be-tested process chains; if none, create 'sample' process chains
    Realization     PHASE 3
    Prior to Upgrade: Install any necessary frontend software    
    Prior to Upgrade: Install any necessary frontend software (end-users)    
    Prior to Upgrade: Complete any extraction - suspend V3 and scheduled BW jobs    
    Create Sandbox    
    Connect Sandbox to external systems (connections/BDLS)    
    Pre-upgrade steps    
    Perform Upgrade of BW     
    Check connection with R/3/APO    
    Perform system backup    
    "BackEnd Test: Logistics extraction and load, process chains
    u2022 Z_example_CHAIN
    u2022 2lis_03_BF and 2lis_03_BX     Test:
    - Extractor within process chain preferably otherwise mimic behavior (drop index, load, rebuild, aggragate etc)
    - Delta extraction"
    "BackEnd Test: Sales Extraction and load, process chains
    u2022 Z_SALES_1
    u2022 all lis extractors:
    u2022 12_vcitm
    u2022 13_vditm
    u2022 11_vaitm
    u2022 11_vahdr
    u2022 11_v_ssl
    u2022 11_vasti
    u2022 12_vcscl
    u2022 selection of z-extractors
         Test:
    - Extractor within process chain preferably otherwise mimic behavior (drop index, load, rebuild, aggragate etc)
    - Delta extraction"
    "BackEnd Test: Financial Extraction and load, process chains
    u2022 0FI_GL_4
    u2022 NOTE: no further extractors are chosen for testing
    u2022 tbd     Test:
    - Extractor within process chain preferably otherwise mimic behavior (drop index, load, rebuild, aggragate etc)
    - Delta extraction"
    "BackEnd Test: Master Data... Extraction and load, process chains     Test:
    - Extractor within process chain preferably otherwise mimic behavior (drop index, load, rebuild, aggragate etc)
    - Delta extraction"
    "BackEnd Test: Zxxx Extraction and load, process chains
    u2022 Z_TRANSACTION_DATA
    u2022 Zxxx
    u2022 Zyyy     Test:
    - Extractor within process chain preferably otherwise mimic behavior (drop index, load, rebuild, aggragate etc)
    - Delta extraction"
    "BackEnd Test: Plant Maintenance Extraction and load, process chains     Test:
    - Extractor within process chain preferably otherwise mimic behavior (drop index, load, rebuild, aggragate etc)
    - Delta extraction"
    BackEnd Test: PP Extraction and load, process chains     Test: load from Zxxx
    "BackEnd Test: Purchasing Extraction and load, process chains     Test:
    - Extractor within process chain preferably otherwise mimic behavior (drop index, load, rebuild, aggragate etc)
    - Delta extraction"
    "BackEnd Test: APO Extraction and load, process chains
    u2022 ZAPOxxx
    u2022 ZAPOyyy
    u2022 others can be tested, but were not included in de testset     Test:
    - Extractor within process chain preferably otherwise mimic behavior (drop index, load, rebuild, aggragate etc)
    - Delta extraction"
    BackEnd Test: Regular Schedule - Production    
    BackEnd Test: Flat file upload     Test: 2 types from server and PC
    BackEnd Test: Test Admin Workbench: creation of objects, change existing     Test the Admin Workbench: creation of infoobjects, change infoobject, create/change hierarchy, create cube,u2026
    BackEnd Test: Test Bex: creation of queries (kf etc...); change existing     Creation of new query/ change query, create condition, calculated/restricted kf, variables,..
    BackEnd Test: Issue handling    
    Development box issue fixing - on forehand     apply fixes based on issues BWQ
    FrontEnd Test: Run Logistics Reports (log results in Reports tab)     Run reports, execute various navigation steps, make sure to execute each item of menu paths,validate no loss of functionality, report new functionality. Run few managed bookmarks
    FrontEnd Test: Run Sales Reports (log results in Reports tab)     Run reports, execute various navigation steps, make sure to execute each item of menu paths,validate no loss of functionality, report new functionality. Run few managed bookmarks
    FrontEnd Test: Run Supply Chain Reports (log results in Reports tab)     Run reports, execute various navigation steps, make sure to execute each item of menu paths,validate no loss of functionality, report new functionality. Run few managed bookmarks
    FrontEnd Test: Run Financial Reports (log results in Reports tab)     Run reports, execute various navigation steps, make sure to execute each item of menu paths,validate no loss of functionality, report new functionality. Run few managed bookmarks
    FrontEnd Test: Run Controling Reports     Run reports, execute various navigation steps, make sure to execute each item of menu paths,validate no loss of functionality, report new functionality. Run few managed bookmarks
    FrontEnd Test: Run Master Data and Special Reports (log results in Reports tab)     Run reports, execute various navigation steps, make sure to execute each item of menu paths,validate no loss of functionality, report new functionality. Run few managed bookmarks
    FrontEnd Test: Run PM Reports (log results in Reports tab)     Run reports, execute various navigation steps, make sure to execute each item of menu paths,validate no loss of functionality, report new functionality. Run few managed bookmarks
    FrontEnd Test: Run Purchasing Reports (log results in Reports tab)     Run reports, execute various navigation steps, make sure to execute each item of menu paths,validate no loss of functionality, report new functionality. Run few managed bookmarks
    FrontEnd Test: Run APO Reports (log results in Reports tab)     Run reports, execute various navigation steps, make sure to execute each item of menu paths,validate no loss of functionality, report new functionality. Run few managed bookmarks
    FrontEnd Test: Issue handling    
    FrontEnd Test: Make before image of reports - Production     Run reports; save of before image to compare once upgraded
    FrontEnd Test: Run All Reports - Production     Run reports, execute various navigation steps, compare with 'before image' taken in prevous step
    Test Security     Test security with various end-user roles/IDS other than BW developer (usually has * ALL)
    Test Portal integration     Test portal with various end-user roles/IDS other than BW developer (usually has * ALL)
    Edited by: A. Nagelhout on Feb 12, 2010 10:50 AM
    Edited by: A. Nagelhout on Feb 12, 2010 10:50 AM
    Edited by: A. Nagelhout on Feb 12, 2010 10:52 AM
    Edited by: A. Nagelhout on Feb 12, 2010 10:52 AM

  • Best way to handle all erros and get performance(OCI)?

    Hi there,
    Im using[b] Oracle Call Interface to execute batch file process. But I have got a problem.
    I set ExecuteBatch with the same number of commit time, i.e: 100, 1000 or just 10, for thats ok.
    ((OraclePreparedStatement)globalStmt).setExecuteBatch(commit);
    And I handle executeUpdate to catch all SQL Exceptions. I made some proposital files with invalid erros but when I handle "executeUpdate" it�s doesn�t get the corrent error line, and puts out another line that is corret.
    ((OraclePreparedStatement)globalStmt).executeUpdate();
    Check in the code I concluded that its always get the same sequence of commit number like the error line. For example, I have between line 1 - 50 a line error, this line is 31, but I set the commit time for 50, its show me that line errror is 50 wherever 31. But if I put setExecuteBatch with '1' so it can handle corret lines erros, but the system performance bring down. What is the best way to handle all erros and keep the perfronace?
    Sorry for my english, I am not native. Thanks all.

    So by doing this, everything will transfer and look exactly the way I have it on the old machine?
    That is correct, if your old machine is Intel based after using MA the new machine will look just like the old machine. Here is information from Apple on MA, I'd recommend looking it over.
    My recommendation is to answer NO when setting up the new machine when it asks "Are you moving from another Mac?" The reason being let you new machine get set up and run for a couple of hours to ensure it's fine. Then launch MA and follow the prompts, it's very easy and if you use a fast connection like FW it should go smoothly.
    Regards,
    Roger

  • Best way to handle source files

    Hi there,
    After some pretty general advise please.
    The company I work for looks after a lot of websites, and one
    of the
    headaches we have is the best way to handle source files. By
    source files I'm
    referring to Photoshop files, Flash .fla files and also other
    none Adobe
    files that relate to a site, not the .html, .asp, .aspx,
    .css, .js etc type
    files.
    Now I'm NOT after a version control system, just a simple way
    to store the
    source files in a location that is separate from the website
    but still to be
    able to have a smooth workflow between the Dreamweaver site
    and it's source
    files.
    At the moment, and I know this is unwise, we have a
    subdirectory within the
    site where we store the source files, and use WebDAV to
    transfer both site
    and source files to and from the server. But I really want to
    separate the
    site from source but still maintain a link between site and
    source...... if
    you see what I mean. I think the upshot is I would like to be
    able to open a
    site within Dreamweaver and instantly be able to access that
    sites source
    files if needed. This method needs to be shared across a
    small team spread
    round the UK.
    I looked at the Repository Subversion version control, but
    like I said I'm
    not after a source control system, plus it appeared to
    conflict with WebDAV
    and Contribute, that some of our clients use to maintain
    content on their
    sites. I also looked at Version Cue which looked promising,
    but can't see a
    clear workflow between Dreamweaver and Version Cue which
    separates site from
    source. I might be missing something.... part of my brain
    perhaps. :)
    Would be grateful for any advice please.
    Cheers,
    @ndyB

    Take a deep breath. Relax. All is fine.
    iDVD does not look at the size of your video file, it looks at the length. iDVD can accomodate up to 2 hours of movie
    iDVD gives you different options depending on the length of your movie. Although I won't agree with your friend about reducing the length of your movie to 15 minutes, if you could trim out a few minutes to get it under an hour that setting in iDVD (Best Performance though the new version may have renamed it) gives you the best quality. Still, any iDVD setting will give you good quality even at 64 minutes
    In FCE export as Quicktime Movie NOT any flavour of Quicktime Conversion. Select chapter markers if you have them. If everything is on one system unchecked the Make Movie Self Contained button. Drop the QT file into iDVD

  • Best way of handling large amounts of data movement

    Hi
    I like to know what is the best way to handle data in the following scenario
    1. We have to create Medical and Rx claims Tables for 36 months of data about 150 million records each - First month (month 1, 2, 3, 4, .......34, 35, 36)
    2. We have to add the DELTA of month two to the 36 month baseline. But the application requirement is ONLY 36 months, even though the current size is 37 months.
    3 Similarly in the 3rd month we will have 38 months, 4th month will have 39 months.
    4. At the end of 4th month - how can I delete the First three months of data from Claim files without affecting the performance which is a 24X7 Online system.
    5. Is there a way to create Partitions of 3 months each and that can be deleted - Delete Partition number 1, If this is possible, then what kind of maintenance activity needs to be done after deleting partition.
    6. Is there any better way of doing the above scenario. What other options do I have.
    7 My goal is to eliminate the initial months data from system as the requirement is ONLY 36 months data.
    Thanks in advance for your suggestion
    sekhar

    Hi,
    You should use table partitioning to keep your data on monthly partitions. Serach on table partitioning for detailed examples.
    Regards

  • Can only access emails through OWA after migration from exchange 2007 to 2013

    can only access emails through OWA after migration from exchange 2007 to 2013, in other words unable to access mails through outlook or from other Applications services.
    needed RCA ... plz help..

    Hi,
    From your description, you can send and receive messages only when you use OWA after migration from Exchange 2007 to Exchange 2013. If I have misunderstood your concern, please let me know.
    In your case, I recommend you create a new test mailbox in your Exchange 2013 and check if you can send and receive messages on Outlook. If yes, it is recommended to create a new profile to solve this issue.
    Hope this can be helpful to you.
    Best regards,
    Amy Wang
    TechNet Community Support

  • Hotkey (Ctrl+B) is not working after migration from Forms 6i to 10g

    Dear Gurus,
    Need your help regarding the following problem:
    After migrating from Forms 6i to 10g (Rel 2), Ctrl+B (used to list the blocks in a form) hotkey is not working anymore (all other hotkeys are working fine). I've checked my FMRWEB.RES and it contains an entry for Ctrl+B:
    66 : 2 : "Ctrl+B" : 70 : "Block Menu"
    Additional Information:
    OS: AIX
    NLS_LANG = UTF8
    fmrweb_utf8.res also contains the same text for Ctrl+B as above.
    I tried to reproduce it on my Windows Machine (with NLS_LANG set to AMERICAN_AMERICA.WE8ISO8859P1) and its not working as well. I don't have right to put move files on AIX server but i can try anything (on my local system) suggested by the Oracle gurus.
    Best Regards,

    Dear Gerd,
    "are you sure, that it is the hotkey"
    Yes, i am dead sure. Since:
    (1) It was working fine in forms 6i
    (2) If you run a form and select Help => Keys, you will see Ctrl+B on the top of list
    The problem is its not taking into account the keys which i am defining in the file (although i am following the same procedure detailed on metalink.oracle.com).
    Regards

Maybe you are looking for

  • Itunes on multiple computers and icloud

    I have itunes 10.4.1 on my home computer and on my work computer (both on same account).  Is it possible to transfer downloads on one computer to the other using icloud?

  • Burn files onto DVD

    Can someone explain how to burn files onto a blank DVD, please?  I have iMac with Lion and DVD-RW.  When I insert the DVD, no icon appears on my desktop.  If I go into the disk utility, the disk shows up there.  Then I click BURN, and a window opens

  • Eek! My cover is blown... Camcorder advice needed...

    Please excuse the bizzarre Thread Title, but i think it sums up my situation perfectly. Basically, i've been using iMovie and iDVD to make some gorgeous home movies reently, and everyone's been doing the usual "Wow! How did you do that?" sort of thin

  • Newbie Table Data Source Question

    Hi, I have a question about table data sources. I can set them up OK but I can't find the answers I'm looking for in the documentation or online. With a table data source, the content column is the one that's indexed right? Are the attribute column s

  • What to do with a i586 processor

    Hello, I got a Gecko Edubook: www.norhtec.com/products/gecko/ I'm not gonna lie, I'm a lazy person, and therefore I love arch linux. Although arch doesn't run on i586. What would you suggest being the easiest way to get this thing up and going, with