Issue in tRFC - missed delta records

Hi All,
There was a load failure few days ago in BI and we can see entries in tRFC. Mistakenly the request status was set to green in BI and so subsequent loads completed successfully. Now we found that the records are missing in BI. When we try to execute LUWs in tRFC, it gives message in BI monitor screen "request that can not be processed since request properly ended".
Is there a way that we can process the data in tRFC or re-extract data?
Cheers,
Sai

HI,
You can go for Full repair request .
It wont affect your delta.
Just run the Full infopasckage with Full repair request and give the selection criteria.
Hope it is helpful.
Thanks,
Saveen Kumar

Similar Messages

  • Missing Delta Records for 2LIS_02_ITM & SCL

    Hi Experts,
    this is how my problem goes.
    i have done my set up table filled on 12th Dec 2010 and from that time onwards the delta were running everyday and filling the DSO and Cube.
    Accidently by some others PC in prod all my delta loads and the setup table load is being deleted except yesterday in PSA for these 2 extractors and now because of some change i have to do a full load to DSO.
    But as the PSA is emply and have only yesterday's request i have, deleted that one as well and done a Init Delta to it and i found out that only the Set up table is comming now and all the deltas in between are missing.
    i have tried a full LOAd to PSA and the result is same.
    How can i get those missing delta records from 12th Dec last year till today with out doing another set up table fill  or Do i have to have fill the set up table again till today and thats the only way? i will set the delta again after that.
    Do we have to have all the user locked for the setup table fill (for Queued Delta type) ? Lot of people says yes you have to and others says no you don't require. i got one white paper and it clearly says no user locking is required. please find the link below. what is the correct way?
    [http://www.sdn.sap.com/irj/scn/go/portal/prtroot/docs/library/uuid/d019f683-eac1-2b10-40a6-cfe48796a4ca?quicklink=index&overridelayout=true]

    Hi,
    As per my knowledge you want load particular period of data try repair request may issue solve.
    Regards
    Sivaraju

  • Missing delta records for the Z extractor

    Hi,
    I have created a Z extractor for the table BUT050 in CRM. I have created a function module and have included the delta logic on CRDAT which is the created date, CRTIM - created time, CHDAT - change date, CHTIM - Change time. I have initialed the delta and started doing the delta loads. I found that there are few missing records in the extractor when compared to the table. I'm attaching my code below.Can anyone please look into it and tell me what the issue is.
    Text removed by moderator
    Thanks
    FUNCTION Z_BI_CC_BUT050.
    *"*"Local Interface:
    *"  IMPORTING
    *"     VALUE(I_REQUNR) TYPE  SRSC_S_IF_SIMPLE-REQUNR
    *"     VALUE(I_DSOURCE) TYPE  SRSC_S_IF_SIMPLE-DSOURCE OPTIONAL
    *"     VALUE(I_MAXSIZE) TYPE  SRSC_S_IF_SIMPLE-MAXSIZE OPTIONAL
    *"     VALUE(I_INITFLAG) TYPE  SRSC_S_IF_SIMPLE-INITFLAG OPTIONAL
    *"     VALUE(I_READ_ONLY) TYPE  SRSC_S_IF_SIMPLE-READONLY OPTIONAL
    *"     VALUE(I_REMOTE_CALL) TYPE  SBIWA_FLAG DEFAULT SBIWA_C_FLAG_OFF
    *"  TABLES
    *"      I_T_SELECT TYPE  SRSC_S_IF_SIMPLE-T_SELECT OPTIONAL
    *"      I_T_FIELDS TYPE  SRSC_S_IF_SIMPLE-T_FIELDS OPTIONAL
    *"      E_T_DATA STRUCTURE  ZBI_BUT050 OPTIONAL
    *"  EXCEPTIONS
    *"      NO_MORE_DATA
    *"      ERROR_PASSED_TO_MESS_HANDLER
    * Auxiliary Selection criteria structure
       DATA: l_s_select TYPE srsc_s_select.
    * Maximum number of lines for DB table
       STATICS: s_s_if TYPE srsc_s_if_simple,
    * counter
               s_counter_datapakid LIKE sy-tabix,
    * cursor
               s_cursor TYPE cursor.
    * Select ranges
       RANGES: l_r_RELNR FOR ZBI_BUT050-RELNR,
               l_r_PARTNER1 FOR ZBI_BUT050-PARTNER1,
               l_r_PARTNER2 FOR ZBI_BUT050-PARTNER2,
               l_r_DATE_TO FOR ZBI_BUT050-DATE_TO,
               l_r_ZZTMSTMP FOR ZBI_BUT050-ZZTMSTMP.
       DATA : startdate LIKE sy-datum,
              starttime LIKE sy-uzeit,
              enddate LIKE sy-datum,
              endtime LIKE sy-uzeit,
              tstamp LIKE tzonref-tstamps,
              timezone type TZONREF-TZONE.
       RANGES: l_r_CRDAT FOR ZBI_BUT050-CRDAT,
                       l_r_CRTIM FOR ZBI_BUT050-CRTIM.
    * Initialization mode (first call by SAPI) or data transfer mode
    * (following calls) ?
       IF i_initflag = sbiwa_c_flag_on.
    * Initialization: check input parameters
    *                 buffer input parameters
    *                 prepare data selection
    * Check DataSource validity
         CASE i_dsource.
           WHEN 'ZCC_MA_BUT050'.
           WHEN OTHERS.
             IF 1 = 2. MESSAGE e009(r3). ENDIF.
    * this is a typical log call. Please write every error message like this
             log_write 'E'                  "message type
                       'R3'                 "message class
                       '009'                "message number
                       i_dsource   "message variable 1
                       ' '.                 "message variable 2
             RAISE error_passed_to_mess_handler.
         ENDCASE.
         APPEND LINES OF i_t_select TO s_s_if-t_select.
    * Fill parameter buffer for data extraction calls
         s_s_if-requnr    = i_requnr.
         s_s_if-dsource = i_dsource.
         s_s_if-maxsize   = i_maxsize.
    * Fill field list table for an optimized select statement
    * (in case that there is no 1:1 relation between InfoSource fields
    * and database table fields this may be far from beeing trivial)
         APPEND LINES OF i_t_fields TO s_s_if-t_fields.
       ELSE.                 "Initialization mode or data extraction ?
    * Data transfer: First Call      OPEN CURSOR + FETCH
    *                Following Calls FETCH only
    * First data package -> OPEN CURSOR
         IF s_counter_datapakid = 0.
    * Fill range tables BW will only pass down simple selection criteria
    * of the type SIGN = 'I' and OPTION = 'EQ' or OPTION = 'BT'.
           LOOP AT s_s_if-t_select INTO l_s_select WHERE fieldnm = 'RELNR'.
             MOVE-CORRESPONDING l_s_select TO l_r_RELNR.
             APPEND l_r_RELNR.
           ENDLOOP.
          LOOP AT s_s_if-t_select INTO l_s_select WHERE fieldnm = 'PARTNER1'.
             MOVE-CORRESPONDING l_s_select TO l_r_PARTNER1.
             APPEND l_r_PARTNER1.
           ENDLOOP.
           LOOP AT s_s_if-t_select INTO l_s_select WHERE fieldnm = 'PARTNER2'.
             MOVE-CORRESPONDING l_s_select TO l_r_PARTNER2.
             APPEND l_r_PARTNER2.
           ENDLOOP.
           LOOP AT s_s_if-t_select INTO l_s_select WHERE fieldnm = 'DATE_TO'.
             MOVE-CORRESPONDING l_s_select TO l_r_DATE_TO.
             APPEND l_r_DATE_TO.
           ENDLOOP.
    * Timestamp is delivered as a selection criterion.
    * Split the timestamp into date and time
          LOOP AT s_s_if-t_select INTO l_s_select WHERE fieldnm = 'ZZTMSTMP'.
             tstamp = l_s_select-low.
             timezone = 'EST'.
             CONVERT TIME STAMP tstamp TIME ZONE timezone
              INTO DATE startdate TIME starttime.
             tstamp = l_s_select-high.
             CONVERT TIME STAMP tstamp TIME ZONE timezone
              INTO DATE enddate TIME endtime.
             l_r_CRDAT-low = startdate.
             l_r_CRDAT-sign = l_s_select-sign.
             l_r_CRDAT-option = l_s_select-option.
             l_r_CRDAT-high = enddate.
             APPEND l_r_CRDAT.
             l_r_CRTIM-low = starttime.
             l_r_CRTIM-sign = l_s_select-sign.
             l_r_CRTIM-option = l_s_select-option.
             l_r_CRTIM-high = endtime.
             APPEND l_r_CRTIM.
           ENDLOOP.
    * Determine number of database records to be read per FETCH statement
    * from input parameter I_MAXSIZE. If there is a one to one relation
    * between DataSource table lines and database entries, this is trivial.
    * In other cases, it may be impossible and some estimated value has to
    * be determined.
           OPEN CURSOR WITH HOLD s_cursor FOR
    * Use the l_r_erdat and l_r_erfzeit for both creation and change selections
    * This way we can pick up both the creations and changes in a given time period.
           SELECT * FROM BUT050
                  WHERE RELNR IN l_r_RELNR
                   AND PARTNER1 IN l_r_PARTNER1
                   AND PARTNER2 IN l_r_PARTNER2
                   AND DATE_TO IN l_r_DATE_TO
                   AND ( CRDAT >= startdate AND ( CRTIM >= starttime OR ( CRDAT <= enddate AND CRTIM <= endtime ) ) )
                   OR ( CHDAT >= startdate AND (  CHTIM >= starttime OR ( CHDAT <= enddate AND CHTIM <= endtime ) ) ).
         ENDIF.
         "First data package ?
    * Fetch records into interface table.
    *   named E_T_'Name of extract structure'.
         FETCH NEXT CURSOR s_cursor
                    APPENDING CORRESPONDING FIELDS
                    OF TABLE e_t_data
                    PACKAGE SIZE s_s_if-maxsize.
         IF sy-subrc <> 0.
           CLOSE CURSOR s_cursor.
           RAISE no_more_data.
         ENDIF.
         s_counter_datapakid = s_counter_datapakid + 1.
       ENDIF.              "Initialization mode or data extraction ?
    ENDFUNCTION.
    Message was edited by: Matthew Billingham

    Hi,
    As per my knowledge you want load particular period of data try repair request may issue solve.
    Regards
    Sivaraju

  • How to recover the missing delta records from R/3

    Hi Experts,
    I am Beginner in BI.
    User has created a ticket stating that, he found that there are only three invoice numbers pulled from R/3 (2lis_13_vaitm) to cube for date on 24.05.2010 where has in r/3 there are 372 invoice numbers are available.  
    i was suggested to recover the data for 24th may'10.
    Please tell me detailed steps to recover the data for one day. Invoice data source will feed the data to DSO and then CUBE.
    Thanks in Advance.
    Latha.

    Hi,
      Please search in forum for how to do a repair full load. You have to fill the setup tables for 24th of may and then do a repair full load.Also check is there any restriction which prevented the invoices from loading.
    Regards,
    Raghavendra.
    Edited by: Raghavendra Padmanaban on Sep 23, 2010 1:42 PM
    Edited by: Raghavendra Padmanaban on Sep 23, 2010 1:43 PM

  • Missing V2 records using Unserialized V3 deltas in Logistics

    We have been experiencing missing Deltas in our Material Movement transactions feeding BW. This always occurs between 5:30AM and 6:00AM. When we examine the intermediate queues we notice the Deltas are not present in the V2 queue. I have read several articles describing the process by which SAP moves Deltas through its arthitecture in order to get Deltas from ECC all the way to the Delta Queue in BW using the Unserialized V3 update method of extraction. All the articles state very clearly that if the V2 update fails, then the transactions are not going to make it through to the Delta queue - that makes sense.
    What none of the articles explain, and what doesn't make sense, is why the V2 updates would fail in the 1st place. So...that is my question - Under what circumstances would the system not complete a V2 update? As a result of messing around with this problem for a couple of weeks and not getting anywhere, we have abandoned the Unserialized V3 Delta process in favor of using the Direct Delta technique. The up side is we will be getting all the Deltas into BW. The down side is that we have not solved our problem - we've only circumvented it, and the next time we attempt to implement the Unserialized V3 update technique, we most likely will go through a similar period of frustration.
    I'm NOT looking for info on "where to look". I'm looking for "concepts" - whiy would a V2 udpate fail? If I know the answer to that question, then the answer to "where to look" should be obvious.
    Thanks in advance for someone's help.

    Hello,
    Here are some useful tips (in my opinion) to this problem:
    - Refer to transaction SM13 to view the 'V2' requests which aren't made into 'V3' Queue. Choose the status option 'V2 Executed', user as '*"  and select the appropriate time. This transaction would provide all the V2 requests which wern't updated. It's also possible to re-post the failed updates using the "Repeat update" option
    - Normally there will be a dump recorded to respective failed update and this can be check using the transaction ST22. If you analyze the problem then you might be able to find the root cause of this failure.
    We are using the 'Queued' delta method for application '03' and this method is working fine and never missed any documents. May be it's worth checking this update method.
    Hope the above tips will help you.
    Cheers
    Bala Koppuravuri

  • Delivery Delta Records missing

    Hi Friends !
    We have missd the delta records for 4 days. Only way I know to run the setup table again & get all the records till date. Then schedule V3 jobs & next day onwards delta records will be pulled. Do we have any other way to get those missed records? Please advise.
    With Regards
    Rekha

    Hi Rekha,
    Please follow the below steps in R3 side
    goto RSA3-->give DS name press enter
    below you will get selection screen
    give you delivery numbers as selections
    then execute, if you get records here then definitely you should get records when you run the IP.
    Hope this helps
    Regards,
    Venkatesh.

  • How to fill set up tables with out missing the delta records

    Hi,
    I would like fill set up tables in the productioon system of apllication of logistics.
    Can you please guide me how do we perform.?
    What are points to be considered?
    Because,when i start the filling  set up table by 10.AM if there are any posting at 10:05,10:06....like that
    how can collect them i.e will i miss any records in second delta run?What setps to be taken care?
    Thanks in advance
    Naresh.

    Hi.
    You can fill the set-up tables during normal operation hours ,if you load the data into ODS and the update queue is 'Queued delta' .Downtime is needed to avoid the duplicates .But if you use  'Direct delta' you miss the delta documents. Hence it is better to go for downtime approach for this case.
    Initially your delta records will be stored in the extraction queue and then when you run the collective job, records will be moved into delta queue. You can run the collective job (LBWE) anytime after the init run.If you need a daily delta ,then schedule this job before the delta loading. You can schedule this job either hrly or daily .This will move your records into delta queue. At the time of delta loading ,all your delta queue records will be moved into BW .
    Thanks.

  • Problem in getting Generic Delta records to BW

    Hi BW Gurus,
    I have got one issues with which I have been struggling a lot for several days . i.e
    I am extracting data from R/3 using Generic Extractor (View) from CATSCO and CATSDB. At the time of delta, I tried using Personal  No with Time Stamp giving Upper limit as 1800 seconds. I executed Infopackage in BW immediately. But it didn't work out. So plz guide me how I can get Delta Records to BW. Or should I do any  necessary factors apart from these in Generic Delta Screen (RSO2).
    It is very urgent.
    I will be thankful for solving this issue.

    Hey I am "Intros" again,,
    I am sorry man for giving the wrong T-code mistakenly in my last reply..
    To solve the proble...the whole process is same what I told u before..
    But... you got to go to T-Code :  BD87   , then execute the IDocs manually to bring them in to BW
    identify your IDocs which are missing in Monitor screen, then goto BD87 and select those IDocs and click execute... )you can search the IDoc's based on selection conditions)
    I hope this will help u...
    cheers man..
    ---Intros

  • Transfer (IDocs and TRFC ) Missing Messages

    Hi,
       I have taken INIT load and records were loaded successfully in the target system.Now when i run delta , the status goes Green but 0 records are loaded ,and also there is error as :
    Requests (messages): Everything OK 
    Data request arranged
    Confirmed with: OK
    Extraction (messages): Everything OK
    Data request received
    Data selection scheduled
    No data available, data selection ended
    Transfer (IDocs and TRFC): Missing messages
    Processing (data packet): No data
    Subseq. processing (messages) : No errors
    Process Chains : Everything OK
    Other (messages): Everything OK
    I have checked extractor in source system and it has records in it.
    Edited by: Bhavna_m on Aug 26, 2011 12:57 PM

    Hi,
    Transact RFC error 
    tRFC Error - status running Yellow for long time (Transact RFC will be enabled in Status tab in RSMO).
    Step 1:  Goto Details, Status get the IDoc number,and go to BD87 in R/3,place the cursor in the RED IDoc entroes in tRFC
    queue thats under outbound processing and click on display the IDOC which is on the menu bar.
    Step 2: In the next screen click on Display tRFC calls (will take you to SM58 particular TRFC call)
    place the cursor on the particular Transaction ID and go to EDIT in the menu bar --> press 'Execute LUW'
    (Display tRFC calls (will take you to SM58 particular TRFC call) ---> select the TrasnID ---> EDIT ---> Execute LUW)
    Rather than going to SM58 and executing LUW directly it is safer to go through BD87 giving the IDOC name as it will take you
    to the particular TRFC request for that Idoc.
    OR
    Go into the JOB Overview of the Load there you should be able to find the Data Package ID.
    (For this in RSMO Screen> Environment> there is a option for Job overview.)
    This Data Package TID is  Transaction ID in SM58.
    OR
    SM58 > Give * / user name or background  (Aleremote) user name and execute.It will show you all the pending TRFC with
    Transaction ID.
    In the Status Text column you can see two status
    Transation Recorded and Transaction Executing
    Don't disturb, if the status is second one Transaction Executing. If the status is first one (Transation Recorded) manually
    execute the "Execute LUWs"
    OR
    Directly go to SM58 > Give * / user name or background  (Aleremote) user name and execute. It will show TRFCs to be executed
    for that user. Find the particular TRFC (SM37 > Req name > TID from data packet with sysfail).select the TrasnID (SM58) --->
    EDIT ---> Execute LUW
    Thanks,
    JituK

  • How SAP determines Delta records?

    Dear Experts,
    I have a standard extractor 2LIS_02_ITM. It works like this.
    1. when a sales order is saved, a customized program creates a new PO and updates a field in the PO using direct update, the abap 'update' command.
    2. in BI, this record created is loaded as delta.
    3. We notice that sometimes this when delta loaded, this field is blank and sometimes its not.
    A. We suspect the delta did not capture program-created PO if this field value is populated as blank by the program.
    B. If the program-created PO is opened and edited by user and then saved, SAP would generate a Delta record?
    If this delta is created and BI reloads delta, this delta would be loaded into BI and the report would show the field value.
    So, I need to check with experts here if using customized program to create a PO and update a field in it at the same time, would this be captured by the Delta mechanism?
    Normally, if manually created, the PO would, i assume, be captured in the delta queue. But what about using program?
    Please advise how delta is or can be affected by programs and how exactly is delta determined in this case. This standard extractor is set to use delta, i think using numeric field.
    So, I am not sure if its due to R/3 or BW. For BW, its a direct mapping of the updated field. So, I do not think there is an issue on BW side.
    regards
    Pascal

    Dear Debjani,
    On your reply to point 2 about the timing of LIS V3 update run, I like to seek clarification how the timing could cause the problem given that many delta records have successfully been loaded with existing way where a collection run is scheduled to run every around 15 minutes and is able to capture the delta except in some cases which is the problem? How can the field value change be missed due to this?   On your point 3 reply, the PO is created by program and it can be created by user.
    I still require some clarification . Hope you can kindly bear with me.
    1. For most of the delta records loaded into BI from R/3, there is no problem with the data. The program-created PO records can be loaded. But for some cases, even when this field is populated by program or EDI, after auto-creation of the PO. This delta is loaded into BI with this field being blank.
    2. This suggests to me, for most cases of the auto-PO creation, the BI Delta mechanism in R/3 is able to handle PO created by automated program in R/3.
    3. So, seems to me, regardless of how the PO is created : user created / by program / by EDI, the delta mechanism seems to work. So, I like to learn more how exactly can TIMING difference (asynchronous) between the creation of PO records and the LIS V3 update delta queue creation batch job have resulted in the field update from not being extracted into the delta queue ?
    Solutions (i think):
    1. Is it possible to Change the current delta queue update extraction method from periodic (every x minutes) V3 update batch run to the Direct / Unserialised V3 update  extraction method as mentioned by Arvind earlier? Would this mean every time PO is created by any means, even by program or field updated by EDI, the delta queue is immediately created to capture the change? This sounds like a solution?
    2. Have the auto-PO creation program Trigger an R/3 event to run the V3 update whenever it has completed an update or PO creation to generate the delta queue. Would this load the r/3 too much if run frequently? Or since frequent run would mean also less delta queue created which means shorter run time?
    Dear Sven,
    would it be good idea to use the program for ME22N to update the particular field After the PO is auto-created by the ME21N program? Would this mean, using these programs will assure that delta is captured?
    Best regards
    Pascal
    Edited by: Pascal Gabin on May 11, 2011 9:24 AM

  • Delta records are not loading from DSO to info cube

    My query is about delta loading from DSO to info cube. (Filter used in selection)
    Delta records are not loading from DSO to Info cube. I have tried all options available in DTP but no luck.
    Selected "Change log" and "Get one request only" and run the DTP, but 0 records got updated in info cube
    Selected "Change log" and "Get all new data request by request", but again 0 records got updated
    Selected "Change log" and "Only get the delta once", in that case all delta records loaded to info cube as it was in DSO and  gave error message "Lock Table Overflow" .
    When I run full load using same filter, data is loading from DSO to info cube.
    Can anyone please help me on this to get delta records from DSO to info cube?
    Thanks,
    Shamma

    Data is loading in case of full load with the same filter, so I don't think filter is an issue.
    When I follow below sequence, I get lock table overflow error;
    1. Full load with active table with or without archive
    2. Then with the same setting if I run init, the final status remains yellow and when I change the status to green manually, it gives lock table overflow error.
    When I chnage the settings of DTP to init run;
    1. Select change log and get only one request, and run the init, It is successfully completed with green status
    2. But when I run the same DTP for delta records, it does not load any data.
    Please help me to resolve this issue.

  • PSA had delta records but while updating to DSO it shows zero records.

    Hi all,
    I have a delta info package which picks up the delta records. The data is successfully extracted from source system and updated in the PSA. But the no of records extracted to the DSO are zero. Couldnt understand what is the problem. Request your valuable inputs on this issue.

    Inside the InfoPackage which you ran, click on the Monitor button or hit F6. In the next screen, open up the folder with the correct run date and select the node - on the right hand side of the screen, click on the tab named Details. Click on No to the subsequent dialog box. Here, open up the last node named "Processing (data packet)" then open up node named "Data Package". Open up all nodes under this one and you'll probably find that either there is a custom start/end routine in the Transfer Rules or in the Update Rules which is filtering out the records from the PSA from updating the target (Cube/ODS). One of these nodes will show you the exact count of how many records came into PSA and how many were passed on to the data target to update it and the reason behind it, for e.g. "Number of data records changed form 16 to 2 in start routine".
    Hope this helps.
    Cheers,
    Sougata.

  • How to reprocess delta records during subsequent update of data targets

    Hi,
    I have an ODS which sends in delta records to a cube above it. Now there is a data load, happening monthy from a flat file, which was successful upto the ODS, but failed in the cube due to one invalid record. How do I reprocess this data load after correction of the invalid records in the source flat file?
    The method I currently follow is to delete the failed request from the cube and then do a selective delete of records from the ODS and then reload the last data after correction.
    Is there a better way to just reprocess the last load from the flat file?
    Thanks & regards,
    Nikhil

    Hi Bhanu,
    Thanks for the response. I will explain a bit more in detail. Suppose there are 5 records coming in from the flat file - now 4 of those records are correct and one has invalid data.
    All 5 records get loaded to the ODS successfully but in the cube we have a function module which works on the data and aborts the entire data package if it finds even one invalid record. So nothing gets loaded into the cube at all.
    Now if we correct the one invalid record and then reload into the ODS - it will just send this corrected record over to the cube and we will miss out on the other 4 records which were originally correct since the ODS does not detect them as changed.
    If we delete the request from the ODS (rather than doing the selective delete) then it disables the delta in the ODS asnd we have to reinitialise everything - which we cannot afford to do in a production environment.
    Hope the problem is a bit more clear now.
    I was looking for some way to just resend all the five records - with the one corrected record to the cube without going through this selective delete process sinc eits too cumbersome and we run the risk of messing up with actual data ina  production environment.
    Thanks for the help,
    regards,
    Nikhil

  • Disk Utility and fsck report 'missing thread records' and can't repair.

    Hello,
    What first made me realise there was a problem with my iMac was that when I tried to update a application it failed. I then manually download the .dmg of the update and tried replaced the app (Transmission 1.82 being that app in question). This failed, OS X claiming that it could not replace the older version (1.81). I then tried copying the file to the desktop where OS X claimed that there was already a copy of the application and so left a corrupted file on my desktop. This also happened when i attempted to move and copy a couple of pages documents and movie files around. I decided the best thing to do at this point was to reboot my computer; OS X decided it needed to install a couple of security updates at this point. This installation failed.
    My iMac booted up, but only after a progress bar has appeared under the Apple logo and spinner and has done its thing (From what I've read this means that OS X is trying its best to resolve some sort of errors). This happens upon every boot.
    This made me think that something must be going on with the file system of my Mac so I ran all the tests in Disk Utility. Permissions repair ran fine and fixed a few issues, but when I verified the disk this is what it reported:
    "Verifying volume “Macintosh HD”
    Performing live verification.
    Checking Journaled HFS Plus volume.
    Checking extents overflow file.
    Checking catalog file.
    Missing thread record (id = 2045563)
    Missing thread record (id = 2096282)
    Keys out of order
    The volume Macintosh HD was found corrupt and needs to be repaired.
    Error: This disk needs to be repaired. Start up your computer with another disk (such as your Mac OS X installation disc), and then use Disk Utility to repair this disk."
    I booted up from the DVD and tried to repair; this failed. Next I tried using fsck in single user mode; this also failed. Upon rebooting again and waiting for the progress bar to fill I was able to move files around and update Transmission without issue. Disk Utility still reports the same error.
    If I boot into Windows Vista everything works perfectly; as of yet I haven't got around to installing the new Boot Camp update for Windows 7 support incase that changes anything. When running the diagnostic tool on the Windows Vista DVD it reports that everything is fine for Windows.
    I really am at a loss for what to do next. All my data is backed up by Time Machine onto a Time Capsule as well as manually onto a FW external. I also made a clone of the Windows partition using the free app Winclone. Does anyone know anything I can do to fix this problem?
    Thanks so much,
    Sam.

    I have been experiencing exactly the same problem here with my MacPro. Apple replaced the hard disk because it was reporting bad sectors. However even with the replacement hard disk I'm experiencing the "missing thread" issue again. I'm guessing its either a software issue that keeps recurring or a fundamental hardware issue. Is the disk controller part of the hard disk or the motherboard? I've never had hard disk problems like this before.
    Here's my Disk Utility report from earlier today:
    Checking Journaled HFS Plus volume.
    Checking extents overflow file.
    Checking catalog file.
    Missing thread record (id = 1067066)
    Missing thread record (id = 1111528)
    Missing thread record (id = 1196845)
    Missing thread record (id = 1200621)
    Missing thread record (id = 1260154)
    Missing thread record (id = 1277166)
    Missing thread record (id = 1285010)
    Missing thread record (id = 1297257)
    Missing thread record (id = 1316679)
    Missing thread record (id = 1437800)
    Incorrect number of thread records
    Checking multi-linked files.
    Checking catalog hierarchy.
    Invalid volume directory count
    (It should be 155085 instead of 155095)
    Checking extended attributes file.
    Checking volume bitmap.
    Checking volume information.
    Repairing volume.
    Missing directory record (id = 1437800)
    Missing directory record (id = 1316679)
    Missing directory record (id = 1297257)
    Missing directory record (id = 1285010)
    Missing directory record (id = 1277166)
    Missing directory record (id = 1260154)
    Missing directory record (id = 1200621)
    Missing directory record (id = 1196845)
    Missing directory record (id = 1111528)
    Missing directory record (id = 1067066)
    Look for missing items in lost+found directory.
    Rechecking volume.
    Checking Journaled HFS Plus volume.
    Checking extents overflow file.
    Checking catalog file.
    Checking multi-linked files.
    Checking catalog hierarchy.
    Checking extended attributes file.
    Checking volume bitmap.
    Checking volume information.
    Invalid volume directory count
    (It should be 155106 instead of 155096)
    Repairing volume.
    Rechecking volume.
    Checking Journaled HFS Plus volume.
    Checking extents overflow file.
    Checking catalog file.
    Checking multi-linked files.
    Checking catalog hierarchy.
    Checking extended attributes file.
    Checking volume bitmap.
    Checking volume information.
    The volume MacHD was repaired successfully.

  • IDOC in 51 status.Essential transfer parameters are missing in record

    Hi All,
    Essential transfer parameters are missing in record: 8701 000001
        Message no. VL561
    I am not understanding what is missing in the IDOC.
    Please help me in resolving the issue.
    Thanks,
    Forum

    hi,
    Do check whether the below points are mentioned or not...
    Outbound delivery
    Shipping point
    Sales organization
    Distribution channel
    Division
    Material number
    Delivery quantity
    Ship-to party
    Base unit of measure
    Inbound Delivery
    Base Unit of Measure
    Conversion factors for converting base and sales units
    Goods Receiving Point
    If it is given here in the doc, then check whether these are getting saved into the table or not..
    Tables are :
    LIKP, LIPS, VTTP...
    Hope it helps..
    Regards
    Priyanka.P

Maybe you are looking for

  • ITunes could not connect to this iPhone because an unknown error occurred (0xE80007F)

    Recieved this error while syncing the IPod classic, however, the IPhone was not connected to the computer, was no where near the computer. In addition, I had earlier in the day the following error message as well: ITunes could not connect to the IPon

  • Input/output sound for MacBook Pro 10.6.8

    The sound is not working properly for my narrated powerpoint presentation. I have recorded it on my MacBook Pro 10.6.8 and I get some sound but it sounds like feedback or something (like a scary phantom voice). I am using the internal microphone and

  • 6.01 Degraded Quality Of Files Compared To Adobe Reader

    I have a wide format scanner, an HP Designjet 4200(815mfp), that can save scans in .pdf format. When I click on the properties tab of one of these .pdf's, this is the info: Application: CTX PDF Producer: PDFlib 4.0.3(Win32) PDF Version: 1.3. I am usi

  • Manual Reorder point planning stock is not showing in MD04

    HI,    Would any one help me when i plan for reorder point planing the details furnished below. Mrp type: V1 Lot size :HB ROP : 30 Max stock level :1000 in Material master settings.But when i run the MRP using MD03 system is not showing available sto

  • Long fields substitution

    Hi, after i read some records in my table, the value of the long fields of them are changed by the value of other long fields of other records and viceversa.. can somebody help me? thanks null