Changing date on a solaris system.

Hi guys,
one easy question.
I need to change the time/date on several solaris servers running oracle 9.
My question is if I need to stop the databases after doing the change or not.
Is it going to affect in any way the normal working of the oracle system?
Thanks in advance.
MLN

There is no need to stop Oracle-- it will immediately start picking up the new date.
Whether changing the time & date will have a negative impact on whatever applications you are running, though, is something you would have to determine. If your applications are using SYSDATE to populate rows, and the application makes certain assumptions about the distribution of records over time (i.e. the most recent row has the latest SYSDATE, there is no more than x seconds between subsequent rows, etc) changing the OS time & date could foul up that logic. Generally, that's more likely to happen if you are setting the date back, but can happen if you advance the clock as well.
Justin

Similar Messages

  • Last changed date different between two systems after transport

    Hi, first of all, I am not from basis background and just want to make sure whether is this something correct or not.
    I found that the "Last changed date" and "username" are not same between the original system and the target system. Is this something configurable?
    And the "username" field is totally empty. However, the program code's version is identical in both systems.

    I think I misunderstood your question.  You are referring to the system Transports not the connections between systems.  If the Transports are not transferring all the contents, you really should talk to your basis folks and tell them which items are not being reproduced correctly.
    Regards,
    Mike

  • Error in the Moment of Change Date ESS Address

    Good Afternoon,
    I have the following issue,
    When the employee tried of change date of address the system display a message of Error:
    ! You have no authorization to Insert!
    The other issue is that the system not update the field Municipality in the moment of select the field of the listbox Region, ie the listbox aren't dynamic.
    Someone know of some note?
    Regards,
    Wilman

    To Begin with its suggested that you either use a standard Employee role, or copy of Standard role. and then you can remove unwanted authorizations from them.
    refer
    helps[help|http://help.sap.com/saphelp_erp60_sp/helpdata/en/99/76b6c6eb2711d1899d0000e8216659/frameset.htm]
    Hope this helps.
    Rgds

  • Not able to change the data of test data containers in production system

    Dear All,
    We have created eCATT scripts in Development SolMan System and moved the transports to Production SolMan System.  Customer wants to change the data at Test data containers and run the scripts in production system but we are not able to edit the data. 
    May be the reason is SCC4 transaction code has set the below option.
    Changes and transports for client-specific Objects
    u2022 No changes allowed
    Customer doesnu2019t want to change the above option and wanted to change the test data containers to give different datau2019s and run the eCATT scripts.
    Could you please let me know the solution for this?
    Your help is really appreciated.
    Thanks,
    Mahendra

    eCatt has the feature where you don't need to transport the scripts or test configuration to our target system. We can keep all our scripts and test data in Solman and run this script any other system in your landscape using the System data container and target system.
    Maintain the production as one of the target system in System container in Solman and point that system while running the script. Change the test data in Solman to run this script.
    Let me know if you need more information
    thanks
    Venkat

  • How to change data in decision table using excel file in operating system?

    Hello, BRF Experts
    I now brf + for the first time in the current project is being developed.
    Now, project has been facing a big problem.
    Develop system that was developed in the history of decision table in the operating system can be used to update excel file, etc. I do not know whether.
    So, I was tested, but can not change data in decision table.
    [ error message : system is not change allowed ].
    The operating system itself to the current system setting is immutable and check it
    : changes and transports for client-specific object -> no changed allowed
    : cross-client object changes -> no changes to repository and cross-client customizing object.
    Decision table in the operating system can update the details if I want to know what to do.
    Maintenance view can be modified in the operating system as a decision table, I wonder if you can create a view.
    Edited by: Sung-Wook Ha on Jan 13, 2012 2:02 PM

    Hi ,
    Exchanging data with Microsoft Excel is done with the help of the XML-based file format (file extension .XSLX) introduced with Microsoft Excel 2007. If you are using an older version of Microsoft Office, you can download the Microsoft Office Compatibility Pack to upgrade the supported file formats from the Microsoft Download Center ([Exchanging Decision Table Data With Microsoft Excel |http://www.microsoft.com/downloads/details.aspx?familyid=941b3470-3ae9-4aee-8f43-c6bb74cd1466]). Once you have downloaded the Compatibility Pack, you have to install it on every workstation that has an older version of Microsoft Office installed.
    Please refer the link-
    [BRFplus-Data exchange with MS Office 2007|http://help.sap.com/erp2005_ehp_05/helpdata/en/b8/e03486d1df48c6baa3c4354f5198ee/frameset.htm]
    Hope this helps.
    Thanks,
    Ajay

  • Change data capture from DRM to Target system

    Hi All,
    I have a situation where client wants Hyperion DRM to be single source of truth and pass Hierarchy information from DRM to subscribing target systems. There are 3 target systems and the database is MS SQL server.
    Now in the export profile ,I am using database export where I have mapped the DRM node/properties to target table's columns. For the initial load it is a new record in the database and the data flows as it is a 1:1 mapping, however if few of the property value changes and I wan to update only the corresponding columns what will the architecture for push the change data in to target.
    Can we achieve this without have any staging tables and capture the change data ?
    Do need to use ODI ?
    Does DRM has any work around without putting extra effort hrs by building CDC tables ?
    I will really appreciate your suggestions and comments on this......

    Thanks for the quick reply. However I was trying to resolve the Change data capture in Hierarcy through as" Database Table" but it seems when we choose the DEVICE option as database in the target tab of export it can only export or insert new records into the tables but cannot update and existing record.
    PLease correct me if my understnding is wrong ?????

  • HT3140 put a date 1970 june 27 . and now the whole system is out of control i cant get to internet cannot change date and time  always when i opened system preferences stays in a waiting mode but doesnt show any opcion don t know what to do and i just bou

    put a date 1970 june 27 . and now the whole system is out of control i cant get to internet cannot change date and time  always when i opened system preferences stays in a waiting mode but doesnt show any opcion don t know what to do and i just bought this computer

    It may well be a driver issue but considering the loading: Xorg trying to load different drivers is somewhat normal. @cafe you can see that these are autoconfigured by X above the loading in the log:
    Jul 20 19:48:24 arch gdm-Xorg-:0[216]: (==) Matched intel as autoconfigured driver 0
    Jul 20 19:48:24 arch gdm-Xorg-:0[216]: (==) Matched intel as autoconfigured driver 1
    Jul 20 19:48:24 arch gdm-Xorg-:0[216]: (==) Matched modesetting as autoconfigured driver 2
    Jul 20 19:48:24 arch gdm-Xorg-:0[216]: (==) Matched fbdev as autoconfigured driver 3
    Jul 20 19:48:24 arch gdm-Xorg-:0[216]: (==) Matched vesa as autoconfigured driver 4
    As long as one of the drivers gets loaded, which it apparently does the failing of the others is fine.
    Was there an update maybe ?

  • Change Data Source not showing tables (in the taget drop down) that begin with "SYSTEM"

    I am trying to change the data source to a database table that is named someschema.SYSTEM_XXX_XXX but none of my tables that are named SYSTEM_.......  are in the combo box list of the taget tables (Change Data Source) dialog.   I have tried this on several database with the same results.   Are you filtering out tables that begin with system ?

    <p>This issue has been resolved in the latest hot fix for CR4E.   You can download the latest version of the hot fix by   following these steps:</p> <ol><li>Go under the <strong>Help</strong> menu</li><li>Select the option <strong>Software Updates -> Manage Configuration</strong></li><li>Expand the tree until you see <strong>Crystal Reports for Eclipse 1.0.1.v671</strong> (or similar)</li><li>Highlight this entry and select the <strong>Scan for Updates</strong> option in the window on the right.</li></ol> <p>This will allow you to update to the latest version which is   currently <strong>1.0.2.v795</strong> </p> <p>Please let me know if this fix does not resolve the issue   outlined in this thread. </p><p> </p><p>Regards,<br />Sean Johnson (CR4E Product Manager) <br /><br /> <a href="http://www.eclipseplugincentral.com/Web_Links-index-req-ratelink-lid-639.html">Rate this plugin @ Eclipse Plugin Central</a>          </p>

  • OK to use fdisk/100% "SOLARIS System" partition for RAID6 Virtual Drive?

    Solaris newb, here - I am configuring an x4270 with 16 135 GB drives. Basic approach is
    D0, D1: RAID 1 (Boot volume, Solaris, Oracle Software)
    D2-D13: RAID 6 (Oracle dB files)
    D14, D15: global spares
    After configuring the RAID's w/WebBIOS Utility, I am now trying to format/partition the RAID 6 Virtual Drive, which shows up as 1.327 TB 'Optimal' in the MegaRAID Storage Manager. After hunting around the ether for advice on how to do this, I came across http://docs.oracle.com/cd/E23824_01/html/821-1459/disksxadd-50.html#disksxadd-54639
    "Creating a Solaris fdisk Partition That Spans the Entire Drive"
    which is painfully simple: after 'format', just do an 'fdisk' and accept the default 100% "SOLARIS System" partition. After doing this, partition>print and prtvtoc show this:
    partition> print
    Current partition table (original):
    Total disk cylinders available: 59125 + 2 (reserved cylinders)
    Part Tag Flag Cylinders Size Blocks
    0 unassigned wm 0 0 (0/0/0) 0
    1 unassigned wm 0 0 (0/0/0) 0
    2 backup wu 0 - 59124 1.33TB (59125/0/0) 2849529375
    3 unassigned wm 0 0 (0/0/0) 0
    4 unassigned wm 0 0 (0/0/0) 0
    5 unassigned wm 0 0 (0/0/0) 0
    6 unassigned wm 0 0 (0/0/0) 0
    7 unassigned wm 0 0 (0/0/0) 0
    8 boot wu 0 - 0 23.53MB (1/0/0) 48195
    9 unassigned wm 0 0 (0/0/0) 0
    # prtvtoc /dev/dsk/c0t1d0s2
    * /dev/dsk/c0t1d0s2 partition map
    * Dimensions:
    * 512 bytes/sector
    * 189 sectors/track
    * 255 tracks/cylinder
    * 48195 sectors/cylinder
    * 59127 cylinders
    * 59125 accessible cylinders
    * Flags:
    * 1: unmountable
    * 10: read-only
    * Unallocated space:
    * First Sector Last
    * Sector Count Sector
    * 48195 2849481180 2849529374
    * First Sector Last
    * Partition Tag Flags Sector Count Sector Mount Directory
    2 5 01 0 2849529375 2849529374
    8 1 01 0 48195 48194
    My question: is there anything inherently wrong with this default partitioning? Database is for OLTP & fairly small (<200 GB), with about 140 GB being LOB images.
    Thanks,
    Barry

    First off, RAID-5 or RAID-6 is fine for database performance unless you have some REALLY strict and REALLY astronomical performance requirements. Requirements that someone with lots of money is willing to pay to meet.
    You're running a single small x86 box with only onboard storage.
    So no, you're not operating in that type of environment.
    Here's what I'd do, based upon a whole lot of experience with Solaris 10 and not so much with Solaris 11, and also assuming this box is going to be around for a good long time as an Oracle DB server:
    1. Don't use SVM for your boot drives. Use the onboard RAID controller to make TWO 2-disk RAID-1 mirrors. Use these for TWO ZFS root pools. Why two? Because if you use live upgrade to patch the OS, you want to create a new boot environment in a separate ZFS pool. If you use live upgrade to create new boot environments in the same ZFS pool, you wind up with a ZFS clone/snapshot hell. If you use two separate root pools, each new boot environment is a pool-to-pool actual copy that gets patched, so there are no ZFS snapshot/clone dependencies between the boot environments. Those snapshot/clone dependencies can cause a lot of problems with full disk drives if you wind up with a string of boot environments, and at best they can be a complete pain in the buttocks to clean up - assuming live upgrade doesn't mess up the clones/snapshots so badly you CAN'T clean them up (yeah, it has been known to do just that...). You do your first install with a ZFS rpool, then create rpool2 on the other mirror. Each time you do an lucreate to create a new boot environment from the current boot environment, create the new boot environment in the rpool that ISN'T the one the current boot environment is located in. That makes for ZERO ZFS dependencies between boot environments (at least in Solaris 10. Although with separate rpools, I don't see how that could change....), and there's no software written that can screw up a dependency that doesn't exist.
    2. Create a third RAID-1 mirror either with the onboard RAID controller or ZFS, Use those two drives for home directories. You do NOT want home directories located on an rpool within a live upgrade boot environment. If you put home directories inside a live upgrade boot environment, 1) that can be a LOT of data that gets copied, 2) if you have to revert back to an old boot environment because the latest OS patches broke something, you'll also revert every user's home directory back.
    3. That leaves you 10 drives for a RAID-6 array for DB data. 8 data and two parity. Perfect. I'd use the onboard RAID controller if it supports RAID-6, otherwise I'd use ZFS and not bother with SVM.
    This also assumes you'd be pretty prompt in replacing any failed disks as there are no global spares. If there would be significant time before you'd even know you had a failed disk (days or weeks), let alone getting them replaced, I'd rethink that. In that case, if there were space I'd probably put home directories in the 10-disk RAID-6 drive, using ZFS to limit how big that ZFS file system could get. Then use the two drives freed up for spares.
    But if you're prompt in recognizing failed drives and getting them replaced, you probably don't need to do that. Although you might want to just for peace of mind if you do have the space in the RAID-6 pool.
    And yes, using four total disks for two OS root ZFS pools seems like overkill. But you'll be happy when four years from now you've had no problems doing OS upgrades when necessary, with minimal downtime needed for patching, and with the ability to revert to a previous OS patch level with a simple "luactivate BENAME; init 6" command.
    If you have two or more of these machines set up like that in a cluster with Oracle data on shared storage you could then do OS patching and upgrades with zero database downtime. Use lucreate to make new boot envs on each cluster member, update each new boot env, then do rolling "luactivate BENAME; init 6" reboots on each server, moving on to the next server after the previous one is back and fully operational after its reboot to a new boot environment.

  • Upgrade of Database with Oracle Change Data Capture

    Hello,
    We are upgrading and moving our database to a different server.
    The move is from 10G R1 database on Solaris to 11G R2 on Linux.
    Our database uses Oracle Change Data Capture.
    What is the best way to perform this migration? Unlike in the approach below, ideally, it would be without any manual steps to drop and recreate CDC subscriptions, change tables, etc.
    Thanks.
    Considerations for Exporting and Importing Change Data Capture Objects
    http://docs.oracle.com/cd/B13789_01/server.101/b10736/cdc.htm#i1027532
    Starting in Oracle Databse 10g, Oracle Data Pump is the supported export and import utility for Change Data Capture. Change Data Capture change sources, change sets, change tables, and subscriptions are exported and imported by the Oracle Data Pump expdp and impdp commands with the following restrictions:
    After a Data Pump full database import operation completes for a database containing AutoLog Change Data Capture objects, the following steps must be performed to restore these objects:
    1. The publisher must manually drop the change tables with the SQL DROP TABLE command. This is needed because the tables are imported without the accompanying Change Data Capture metadata.
    2. The publisher must re-create the AutoLog change sources, change sets, and change tables using the appropriate DBMS_CDC_PUBLISH procedures.
    3. Subscribers must re-create their subscriptions to the AutoLog change sets.

    Hello,
    I opened SR with Oracle Support, they are suggesting to perform full database import/export
    Change Data Capture change sources, change sets, change tables, and subscriptions are exported and imported by the Oracle Data Pump expdp and impdp commands with the following restrictions.
    Change Data Capture objects are exported and imported only as part of full database export and import operations (those in which the expdp and impdb commands specify the FULL=y parameter). Schema-level import and export operations include some underlying objects (for example, the table underlying a change table), but not the Change Data Capture metadata needed for change data capture to occur."
    CDC has different implementation methods:
    You may use the below query to determine-
    select SOURCE_NAME, SOURCE_DESCRIPTION, CREATED, SOURCE_TYPE, SOURCE_DATABASE, SOURCE_ENABLED from change_sources;
    – Synchronous CDC: with this implementation method you capture changes
    synchronously on the source database into change tables. This method uses
    internal database triggers to enable CDC. Capturing the change is part of the
    original transaction that introduces the change thus impacting the performance
    of the transaction.
    – Asynchronous Autolog CDC: this implementation method requires a staging
    database separate from the source database. Asynchronous Autolog CDC uses
    the database's redo transport services to transport redo log information from
    the source database to the staging database1. Changes are captured at the
    staging database. The impact to the source system is minimal, but there is some
    latency between the original transaction and the change being captured.
    As suggested in the document-
    Change Data Capture objects are exported and imported only as part of full database export and import operations (those in which the expdp and impdb commands specify the FULL=y parameter). Schema-level import and export
    operations include some underlying objects (for example, the table underlying a change table), but not the Change Data Capture metadata needed for change data capture to occur.
    ■ AutoLog change sources, change sets, and change tables are not supported.
    Starting in Oracle Database 10g, Oracle Data Pump is the supported export and import utility for Change Data Capture.
    Re-Creating AutoLog Change Data Capture Objects After an Import Operation
    http://docs.oracle.com/cd/B19306_01/server.102/b14223/cdc.htm#i1027532
    After a Data Pump full database import operation completes for a database containing AutoLog Change Data Capture objects, the following steps must be performed to restore these objects:
    a. The publisher must manually drop the database objects underlying AutoLog Change Data Capture objects.
    b. The publisher must re-create the AutoLog change sources, change sets, and change tables using the appropriate DBMS_CDC_PUBLISH procedures.
    c. Subscribers must re-create their subscriptions to the AutoLog change sets.

  • Comparing Solaris systems?

    Does anyone have a process that they can recommend for comparing different Solaris systems to look for inconsistencies? We have a 6 node cluster where all 6 nodes should be identical. For now, I want to verify that they have the same patches applied and the same packages installed (including specific versions). I may want to include more as I get further along, but this is the bare minimum to start. I'd also like to be able to have an automated process that can detect changes and alert us as well.
    All systems are Solaris 10 built from the 5/08 OS.
    thanks in advance.

    SAP Suggest to use backup/restore or R3load .
    Yes - because that´s the most uncomplicated way.
    if I want to create a trace file manually and shut down database,sap .
    copy over  data files /origfile/logs from os level to target system and recover the system and install sap.
    There´s no issue with that if you know what you´re doing
    If you don't use sap mentioned procedure  in that case SAP does not provide any support .
    If you have a problem with that copy procedure itself you will be on you own. if the target system is running already you will still get support of course because nobody asks you later, how you did the copy.
    Markus

  • Can we change data in string object.

    Can we change data in string object.

    Saw this hack to access the char[]'s in a String in another thread. Beware that the effects of doing this is possible errors, like incorrect hashCode etc.
    import java.lang.reflect.*;
    public class SharedString {
            public static Constructor stringWrap = null;
            public static String wrap(char[] value, int offset, int length) {
                    try {
                            if (stringWrap == null) {
                                    stringWrap = String.class.getDeclaredConstructor(new Class[] { Integer.TYPE, Integer.TYPE, char[].class });
                                    stringWrap.setAccessible(true);
                            return (String)stringWrap.newInstance(new Object[] { new Integer(offset), new Integer(length), value });
                    catch (java.lang.NoSuchMethodException e) {
                            System.err.println ("NoMethod exception caught: " + e);
                    catch (java.lang.IllegalAccessException e) {
                            System.err.println ("Access exception caught: " + e);
                    catch (java.lang.InstantiationException e) {
                            System.err.println ("Instantiation exception caught: " + e);
                    catch (java.lang.reflect.InvocationTargetException e) {
                            System.err.println ("Invocation exception caught: " + e);
                    return null;
            public static void main(String[] args) {
                    char[] chars = new char[] { 'l', 'e', 'v', 'i', '_', 'h' };
                    String test = SharedString.wrap(chars, 0, chars.length);
                    System.out.println("String test = " + test);
                    chars[0] = 'k';
                    chars[1] = 'a';
                    chars[2] = 'l';
                    chars[3] = 'l';
                    chars[4] = 'a';
                    chars[5] = 'n';
                    System.out.println("String test = " + test);
    } Gil

  • Uploading data from non sap system to sap system

    hi to all experts,
    my requirement is to upload data into the sap system from non sap system . the data is in  a flat file ,but the problem is that  in the flat file all the field lengths and fields are not in order or same as sap system. How to upload the data into sap system ....

    hi,
    the data is in a flat file,but the problem is that in the flat file all the field lengths and fields are not in order or same as sap system.
    If the data is in excel sheet use this FM.
    CALL FUNCTION 'ALSM_EXCEL_TO_INTERNAL_TABLE'
        EXPORTING
          FILENAME                      = P_FILE
          I_BEGIN_COL                   = 1
          I_BEGIN_ROW                   = 1
          I_END_COL                     = 17
          I_END_ROW                     = 3000
        TABLES
          INTERN                        = itab
    EXCEPTIONS
      INCONSISTENT_PARAMETERS       = 1
      UPLOAD_OLE                    = 2
      OTHERS                        = 3
    It will read the data from excel sheet column wise and then store it in ITAB. you can change the order after reading the file.
    In case of Text file use FM  GUI_UPLOAD.
    You can change the order of the fileds after reading it from file, but the length and all you have to make compatible to SAP standard fields.
    Regards,
    Sachin

  • Identifying Changed Dates on Documents in SAP B1

    Greetingsu2026I have an unusual request that is somewhat long, very serious, and hope folks can help. 
    A headquarters customer believes a subsidiary company has u2018made upu2019 dates on invoices for previous billings.  They describe the date changes as u2018radicalu2019 but I have not personally seen these invoices as of yet and am waiting a FAX.  They requested I go into SAP B1 to see if someone has fabricated or made up or changed invoice dates and who it was.  They have some specific large-dollar invoices identified to research.  The person under suspect (who is no longer there) was considered an expert in SAP B1, so maybe she knew all of the different ways to change dates around.  Being an expert means also she might have used some unusual ways to change dates to the ones she wanted.
    Here are some of my ideas to discover this situation, but I hope someone can suggest even more steps because I would like to know every way that this situation can be discovered.
    1. Open the AR Invoices and use Data > Change Log and concentrate on dates.
    2. Run some SQL to validate document dates and system dates.  SELECT T0.DocNum, T0.DocDate, T0.CreateDate, T0.TaxDate, T0.UpdateDate FROM OINV T0 WHERE (T0.DocDate does not equal T0.CreateDate OR T0.TaxDate does not equal T0.CreateDate).  The results would show invoices that maybe had their dates changed.
    3.Run a Document Journal in the Financial Module.  Finance > Financial Reports > Accounting > Document Journal with the u2018Expanded Selectionu2019 of AR Invoices and a date range covering the invoices in question.  Then see if the Transaction Numbers are out of sequence.
    4. Run a Transaction Journal Report Financials > Financial Reports > Accounting > Transaction Journal Report with AR Invoices.  Compare this report to Document Journal.
    5. Run the Utilities functions of Utilities > Check Document Numbering to see if there any anomalies there in the document numbers.
    6. Run the Administration > System Initialization > Document Numbering function to see if a special series has been set up.
    7. Check out the AR Invoice numbering order by running another simple SQL covering a range of the invoices in question. SELECT T0.DocNum, T0.TaxDate, T0.CreateDate, T0.UpdateDate FROM OINV T0 WHERE T0.DocNum > 1583 AND T0.DocNum < 3251 ORDER BY T0.CreateDate.  Then see if the AR Invoice Numbering Sequence is not consistent.
    8. Check out those documents that are connected to each suspect AR Invoice and their associated dates by using the Target and Base Document icons.  For example, if a Deliver Document has dates of January 2007 but the AR Invoice date is September 2006, then someone has played around with dates on at least one of the documents.
    9. Get copies of the complete company database before the alleged changes were made and a company database as of today and compare the two contents.  Only good if the backups are maintained for a long time.
    Here come the questions on the ways I BELIEVE dates can be changed but not sure how to find and need to work on them:
    QUESTION 1 u2013 If a person DTWs invoices up into SAP B1, how can I tell that DTW was used to change the dates, the date of the DTW run, and the dates that were changed?
    QUESTION 2 u2013 If a person uses SQL to update dates on SAP B1, how can I find out that SQL has been used, the date of the SQL run, and the dates that were changed?
    QUESTION 3 u2013 If a person goes out and changes the workstation and/or server Date and Timestamp Properties, how can I find that they did that and the date changes?  You can do this on your PC by just changing the month, date, and year in Microsoftu2019s Date and Timestamp Properties.
    QUESTION 4 u2013 I know SAP has a policy that no SQL can be used to update databases and/or tables and have been told some customers lost support because they did this.  What does SAP do to discover this in their lab?
    QUESTION 5 u2013 Are there any notes in the portals to help me find changed dates?
    QUESTION 6 u2013 Are there any logs SAP B1 produces that can help identify such changes and how do I find and use them?
    QUESTION 7 (and this is big) u2013 Are the other ways dates can be changed on AR Invoices and how would I uncover the way it was done and the needed information of Original Date and Changed Date?
    Many, many thanks for any help provided since this is a serious situation.  If you can make any suggestions it would be greatly appreciatedu2026
    Zal
    Edited by: Zal Parchem on Mar 13, 2009 7:31 AM

    Hi Zal
    If this person was an "Expert" on SAP Business One then the only way they would have changed dates would probably be directly in SQL itself. The best way to determine this would be to hope that they slipped up somewhere and forgot to change a particular date in let's say one specific table.
    For invoices, start with the post date, doc date & create date. Then check each of these to the OJDT/JDT1 tables for any discrepancies. Next check the ADOC/ADO1 table for multiple instances on the invoice. If only 1 instance then check the create date, time & user as well as the aforementioned dates between OINV/INV1 & ADOC/ADO1. If multiple instances, check the dates from 1 instance to the next (This will indicate dates being changed in the system).
    Next check the same dates against OINM table. If any of the dates do not tally then the SQL was changed directly.
    Hope this helps. Let me know if you need further assistance.
    Kind regards
    Peter Juby

  • Loading of transaction data from SAP ECC system failed

    Hi!
    I successfully connected SAP ECC system to SAP BI system.
    The following steps have been executed:
    - user ALEREMOTE with max. authorization
    - RFC destination
    - Distributing Data model
    - Generated Partner profile
    - Maintaining message types in WE20
    Now when I try to load any data from SAP ECC system the loading process in hanging in status "yellow" and never comletes.
    [0FI_AR_4|http://www.file-upload.net/view-1447743/0FI_AR_4.jpg.html]
    The following steps within Load process are yellow:
    Extraction (messages): Missing messages
      Missing message: Request received
      Missing message: Number of sent records
      Missing message: Selection completed
    Transfer (IDocs and TRFC): Missing messages or warnings
      Request IDoc : Application document posted (is green)
      Data Package 1 : arrived in BW ; Processing : 2nd processing step not yet finished
      Info IDoc 1 : sent, not arrived ; Data passed to port OK
      Info IDoc 2 : sent, not arrived ; Data passed to port OK
      Info IDoc 3 : sent, not arrived ; Data passed to port OK
      Info IDoc 4 : sent, not arrived ; Data passed to port OK
    Subseq. processing (messages) : Missing messages
        Missing message: Subseq. processing completed
        DataStore Activation (Change Log) : not yet activated
    Question:
    Can some one give me some technical steps (tcode, report) to solve this problem?
    Thank you very much!
    Holger

    Hi!
    Many thanks for your answer.
    Via BD87 on BW system I detect that all the IDOC's (type: RSRQST) will be received from SAP ECC system.
    Via tcode SM58 I could not detect any entries.
    However the loading status from yesterday is set to "red".
    The errors are:
    Extraction (messages): Missing messages
    Data Package 1 : arrived in BW ; Processing : 2nd processing step not yet finished
    Info IDoc 1 : sent, not arrived ; Data passed to port OK
    Info IDoc 2 : sent, not arrived ; Data passed to port OK
    Can you investigate my issue again?
    Thank you very much!

Maybe you are looking for