Wading into the data questions

Hi everyone
I know that there are plenty other threads with mobile data usage questions, but I am hoping you might be able to shed some light on what seems to be a unique problem.
I almost had heart failure about an hour ago... I checked my mobile data usage in Settings and saw 4GB Sent and 4.2GB Received. So 8.2GB total. That's in about 40 days since I got my iPhone! I think I have about 6 new apps installed with notifications for all of them disabled, Push for Mail (and everything else) is off, so I should know about any data coming or going.
I do check my mail fairly regularly, but don't get anything with attachments that come anywhere near that.
Most of the time I am surfing at home, where I am connected to WiFi, so that should take care of most of the data in any case. I think I've seen about 4 Youtube videos, all of less than 15 minutes, and all of those were while I was at home in any case.
Now, the really odd thing is that my cell network is reporting only about 195MB of data used. (Which I am pleased about as I am on a 300MB data plan). While they do say that some of the bundle information can be delayed, I really don't think that I have used 8GB of data in the last couple of days that they have just ignored and will bring to my attention in a moment.
This really looks like a glitch in my iPhone's software (for info, I have only just noticed this, and haven't installed DataMan to check up on my usage yet) and would be interested to know if anyone else has seen a similar story.
So, to come to my question (since I marked the post as a question):
1) Is this an iPhone error, or could I really have used up 200MB of 3G data a day, without noticing it?
2) If it is the latter, what can I do about it?
Thanks!
John

Okay, so just a little more info...
I have confirmation from my cell provider that I haven't used (or at least been billed for) 8GB of data. What I have used is about 200MB (which is consistent with the excess of the received over the sent that my phone was reporting).
So there you have it... Anyone else seen something odd like this on their i4? I have reset my statistics and will keep an eye on it to see if anything odd happens.
Cheers
John

Similar Messages

  • How to Troubleshoot why data is not moving over into the Data Warehouse after Sql Server Agent Job Run

    Hello,
    Here is my problem:
    Data was imported into the staging area. After resolving some errors and running the job, I got the data to move over to the next area. From there, data should be moving over into the DW.  I have been troubleshooting for hours and cannot reslove this
    issue. I have restarted the sql service services, I have ran a couple packages manually, and the job is running successfully. 
    What are some reasons why data is not getting into the data warehouse? Where should I be looking? 
    Your help is greatly appreciated!!

    Anything is possible.
    So, just to reiterate, running the job manually works, running the scheduled job does not result in errors neither data arriving to the DW, right? And it used to, correct?
    If so, the 1st step would be to examine the configuration(s). But not before you inspect the package. Do you have an ability to export it to a file system and open in BIDS?
    Arthur My Blog

  • How can I put the signature into the data file.

    Hi all,
    I want to put the signature into the data file, mean the data file will be on the top and the signature will be at the buttom.how can i do that?
    Thanks.
    example:-
    dataFile.txt
    Name=...
    Address=...
    Signature=6A07C70E....123FEAB(Hex)

    Thanks for your reply.
    First, i have a keypair (public key and private key).
    I use the private key to initialize the Signature and
    use the Signature to sign a data file ( dataFile.txt) .
    So when i need to comfirm the data file not being modify
    I have to verify the data file with the Signature and the public key.
    I hardcoded the public key in my program so my client (receiver of data)
    need only the Signature file and the data file (2 files).
    So how can i combine the Signature file and the data file into 1 file?

  • The request has not yet been updated into the data targets.

    Request successfully loaded into PSA - start further update
    Diagnosis
    The request has as data target "only PSA". It arrived there successfully.
    System Response
    The request has not yet been updated into the data targets.
    Procedure
    You can now trigger the update manually. The final target is determined in the request tables. You do not have to make any further settings.
    I am getting this error daily. In RSMO screen status tab option called u201CProcess manuallyu201D, so am doing this manually daily. Is there any solution to fix this problem permanently?
    Thanks & Regards,
    Praveen Yagnamurthy.

    Hi Praveen,
    Dont you have Process chains in Production system for loading thid data. If so its better to schedule the PC rather go for manual loading. In Prod system you will not be able to change. You will have the make the changes in the Development Environment & collect these changes in a TR & then move them into Test environment first to make sure it is working fine. And then once it is tested, you can move the TR to Prod sys.
    Hope it helps!
    Regards,
    Pavan

  • Maintain dimension members directly into the data base

    Hi,
    We need to update the dimension members directly into the data base.
    Instead of going into the administration and entering manually.
    I tried just updating the table mbr and dim. but that didn't work.
    Any insight on that?

    To update the dimension members from the backend tables, update the mbr<dimension>> table and process the dimension either from BPC Admin Console or a SSIS BPC Admin Task. The dimension processing will automatically update the dim<<dimension>> table in the backend.
    However, the updated dimension members can not be seen from the frontend in the BPC Admin. To see the updated dimension member sheet in the current client machine, use Export Dimension option in BPC Admin Console.
    To see the updated dimension an all client machines and get reflected in the current view, please follow this URL:
    Re: BPC 7 MS Update dimension (Make_dim task)
    Hope this helps.

  • PowerPivot removed table replicated into the data model on refresh

    Hey
    I am facing the following problem which is absolutely persistent and annoying in nature. I am using a data model created with PowerPivot with connections to three different SQL servers and 25 tables. The bug is that every time I refresh my data, a table
    that was initially added to the data model (and later removed) is replicated in equivalent number of tables that use this connection, which is currently around 12 tables, so I get 12 duplicates of the same unnecessary table on every refresh, named with the
    name of the schema on the SQL server that is followed by a running number. Removing these from the data model on every refresh simply takes too much time on a daily basis. The additional undesired table which is created is what is defined in the SQL query
    of the connection string of this connection.
    I tried to solve this problem by zipping the Excel file and modifying this SQL statement present in the connection string in the connections.xml file in a way that it would not return anything, but this always leads into a corrupted workbook.
    After encountering this I have managed to avoid this problem by making sure that the initial table I'm adding to the data model will remain in the model, but I really don't want to rebuild this current model. Is there any fix for this issue without
    rebuilding the data model?
    Thanks for all the help in advance!

    I am using 64-bit Office 2013 with the recent SP1 upgrade and the included PowerPivot plug-in on a Windows 8.1 machine.
    I always create the connection by using "From SQL server" then fill in the server and log in details and use a ready SQL statement that I have written by using SQL Server Management Studio to import the initial table. For all the preceding tables
    using the same server and schema, I always select it from existing connections and just paste the SQL queries required for any additional tables to the "Write a query that will specify the data to import", rather than using the table import tool.
    I am not sure if it could have any effect that I've been using the type connection "From SQL server" although the database is really running on Azure (there is a possibility to select From Other sources > Microsoft SQL Azure, but the functionality
    seems the same).
    I don't really understand what I should edit from the existing connections section. Just to clarify, apparently the Connection String parameters can be edited from PowerPivot, but not the Command Text which can be viewed from Excel Data > Connections
    > Properties, but is shown in gray with a message "Some properties cannot be changed because this connection was modified using the PowerPivot add-in."
    In the data model I am using two sets of identical tables, but connected to two different servers that are running a database 1:1 in structure. Initially the problem considered only the other of these, but I accidentally reproduced it by, again, editing
    the SQL query in the table that was initially imported when that connection was created (PowerPivot > Design > Table Properties). Now when refreshing I get tables that are structured as in what the "Connection Text" part of the connection would
    produce, which replicated for the equivalent amount of additional tables using the same connection, so now I'm getting 12 tables (the number of tables using the same connection) each named "*schemaname1* Query", "*schemaname1* Query2",
    "*schemaname1* Query3" "*schemaname2* Query" and so on.
    Personally I can definitely see a pattern here. If there isn't a table matching what has been specified in the "Command Text" that was described when the connection was created, then it for some reason runs this query anyway on every table that
    is using the connection.

  • How To UPLOAD a DATA (.DAT) fiel from PC to internal table and then split it into the data different columns

    Hi all,
    I am new to ABAP Development. I need to upload a .DAT file (the file doesn#t have any proper structure-- Please find the .DAT file in the attachment). After uploading the DATA (.DAT) fiel I need to split in into different columns. Refering the attached .DAT fiel the fields in bracets like:
    [Arbeitstag],  [Pecunia], [Mita], [Kunde], [Auftrag] and  [Position] are different fields that need to be arranged in columns in an internal table. this .DAT fiel which I want to upload and then SPLIT it into various fields will will treated as MASTER DATA table for further programming. The program that I had written is as below. Also please refer the attached .DAT table.
    Please if any one could help me. i searched a lot in different forums but couldn't find me  a solution. Also note that the attached fiel is in text (.txt) format here but in real situation the same fiel is in DATA (.DAT) format.
    *& Report  ZDEMO_ZEITERFASSUNG9
    REPORT  ZDEMO_ZEITERFASSUNG9.
    Types: Begin of ttab,
            Rec(1000) type c,
           End of ttab.
    DATA: itab  type table of ttab.
    DATA: wa_tab type ttab.
    DATA: file_str type string.
    Parameters: p_file type localfile.
    At selection-screen on value-request for p_file.
                                           CALL FUNCTION 'KD_GET_FILENAME_ON_F4'
                                            EXPORTING
    *                                          PROGRAM_NAME        = SYST-REPID
    *                                          DYNPRO_NUMBER       = SYST-DYNNR
    *                                          FIELD_NAME          = ' '
                                               STATIC              = 'X'
    *                                          MASK                = ' '
                                             CHANGING
                                               file_name           = p_file.
    *                                        EXCEPTIONS
    *                                          MASK_TOO_LONG       = 1
    *                                          OTHERS              = 2
    Start-of-Selection.
      file_str = P_file.
      CALL FUNCTION 'GUI_UPLOAD'
        EXPORTING
          filename                      = '\\10.10.1.92\Volume_1\_projekte\Zeiterfassung-SAP\BUP_ZEIT.DAT'   " This the file  source address
          FILETYPE                      = 'DAT'
          HAS_FIELD_SEPARATOR           = ';'
    *     HEADER_LENGTH                 = 0
    *     READ_BY_LINE                  = 'X'
    *     DAT_MODE                      = ' '
    *     CODEPAGE                      = ' '
    *     IGNORE_CERR                   = ABAP_TRUE
    *     REPLACEMENT                   = '#'
    *     CHECK_BOM                     = ' '
    *     VIRUS_SCAN_PROFILE            =
    *     NO_AUTH_CHECK                 = ' '
    *   IMPORTING
    *     FILELENGTH                    =
    *     HEADER                        =
        tables
          data_tab                      = itab
       EXCEPTIONS
         FILE_OPEN_ERROR               = 1
         FILE_READ_ERROR               = 2
         NO_BATCH                      = 3
         GUI_REFUSE_FILETRANSFER       = 4
         INVALID_TYPE                  = 5
         NO_AUTHORITY                  = 6
         UNKNOWN_ERROR                 = 7
         BAD_DATA_FORMAT               = 8
         HEADER_NOT_ALLOWED            = 9
         SEPARATOR_NOT_ALLOWED         = 10
         HEADER_TOO_LONG               = 11
         UNKNOWN_DP_ERROR              = 12
         ACCESS_DENIED                 = 13
         DP_OUT_OF_MEMORY              = 14
         DISK_FULL                     = 15
         DP_TIMEOUT                    = 16
         OTHERS                        = 17
      IF sy-subrc <> 0.
    * MESSAGE ID SY-MSGID TYPE SY-MSGTY NUMBER SY-MSGNO
    *         WITH SY-MSGV1 SY-MSGV2 SY-MSGV3 SY-MSGV4.
      ENDIF.
      LOOP at itab into wa_tab.
            WRITE: / wa_tab.
      ENDLOOP.
    I will be grateful to all you experts for ur inputs
    regards
    Chandan Singh

    For every Auftrag, there are multiple Position entries.
    Rest of the blocks don't seems to have any relation.
    So you can check this code to see how internal table lt_str is built whose first 3 fields have data contained in Auftrag, and next 3 fields have Position data. The structure is flat, assuming that every Position record is related to preceding Auftrag.
    Try out this snippet.
    DATA lt_data TYPE TABLE OF string.
    DATA lv_data TYPE string.
    CALL METHOD cl_gui_frontend_services=>gui_upload
      EXPORTING
        filename = 'C:\temp\test.txt'
      CHANGING
        data_tab = lt_data
      EXCEPTIONS
        OTHERS   = 19.
    CHECK sy-subrc EQ 0.
    TYPES:
    BEGIN OF ty_str,
      a1 TYPE string,
      a2 TYPE string,
      a3 TYPE string,
      p1 TYPE string,
      p2 TYPE string,
      p3 TYPE string,
    END OF ty_str.
    DATA: lt_str TYPE TABLE OF ty_str,
          ls_str TYPE ty_str,
          lv_block TYPE string,
          lv_flag TYPE boolean.
    LOOP AT lt_data INTO lv_data.
      CASE lv_data.
        WHEN '[Version]' OR '[StdSatz]' OR '[Arbeitstag]' OR '[Pecunia]'
             OR '[Mita]' OR '[Kunde]' OR '[Auftrag]' OR '[Position]'.
          lv_block = lv_data.
          lv_flag = abap_false.
        WHEN OTHERS.
          lv_flag = abap_true.
      ENDCASE.
      CHECK lv_flag EQ abap_true.
      CASE lv_block.
        WHEN '[Auftrag]'.
          SPLIT lv_data AT ';' INTO ls_str-a1 ls_str-a2 ls_str-a3.
        WHEN '[Position]'.
          SPLIT lv_data AT ';' INTO ls_str-p1 ls_str-p2 ls_str-p3.
          APPEND ls_str TO lt_str.
      ENDCASE.
    ENDLOOP.

  • How to Load Master data text  Description into the Data Target?

    Hi,
    I have a Master data which is an ID, and the description for that ID is added as text description along with the master data.
    The datasource is a flat file load, which has both the fields i.e. ID as well as the description along with many other fields.
    I need to load the flat file without changing it, into the ODS.
    Please suggest a way to load the description of the ID along with the ID into the ODS, and then further into the CUBE.
    This description field should be available, while checking the contents of the ODS as well as the CUBE.

    hi Srinivas,
    Thanks for the quick response.
    I actually have the flat file with the description within it.
    I cannot make any changes to the flat file.
    So, please suggest me accordingly.

  • After 3 years, I got a new iphone5 and ran into the security question issue. Because of the questions apple chose, I have no idea what the answers are. I did the iTunes email thing and waited 36 hours for a response. I

    After waiting the 36 hours, I received an email advising me. To contact a phone support number. After waiting 15 minutes on hold, the helpful person asked me for the answers to the security questions and said that answering them was the only way to reset the questions. I said, WHAT? You don't see any cyclic logic there? Apple (the inovator!) has not foreseen this? I was livid. I was advised I could leave a comment on Apple.com. I think I'll leave my comment on every Apple contract that crosses my desk. I'm IT Systems at a large US university. If anyone from Apple wishes to address this, PM me.

    Thanks! I'm a little leery of attaching a credit card to itunes again because of all the apple database jacking in the past. I had a bunch of unauthorized purchases charged to my account for devices that I never owned. Took me weeks to replace the credit card even though it was never compromised but apple was impotent to fix the issue. On this issue, I finally got the working key to the system. Even though there are BAZILLIONS of ways to verify a person's identity via two forms, using the security questions forgotten as one of them seems to be circular (failed) logic. Seriously, it's like requiring someone's password to reset a forgotten password. My take on the issue fixed by this workflow, that I noticed from the 3 sets of calls and 20 minute wait on each is this.... iTunes was racked by all the fraudulent purchases I (and thousands of others) experienced and people (like me) had to go through all the credit card issues because Apple "would not" fix the issues in iTunes and pursue the fraudulent purchasers (even though Apple had the device IDs!). So we ripped out our credit card info in iTunes. On this security question reset issue, if they can require people to re-enter their credit card info under the guise of "identification ...because your security is so important!) and then take 60 minutes to create a high intensity customer interaction, there is a high probability that the average customer will temporarily forget all about removing their credit card info. It's actually brilliant! ...but very underhanded.... but also very Apple. I don't expect this reply to have a long life-span, but thanks for your response. If I can log into iTunes, I can add any credit card. therefore IF I CAN LOG INTO ITUNES APPLE HAS THE SECOND METHOD ALREADY!  Thanks.

  • Automatic Moving a movie into the Library question please?

    Good Day.
    If I rip one of my own DVD's and have it put directly into the iTunes Move Library, will it come up in the iTunes Movie list window automatically or do I have to do something else?
    Thanks for looking.
    Tom

    Thanks Razor...
    But I was wondering if I had a DVD conversation app write the movie *mp4 etc., right into the "Movie" in the iTunes library would the iTunes app/Window see it when I then launched iTunes? Kinda without going thru the above mentioned step....
    Thanks again.

  • What are the major process to transfer the data from legacy to sap system.

    What are the major process to transfer the data from legacy to sap system using BDC at Real Time only?

    hi,
    BATCH DATA COMMUNICATION
    main methods are:
    1. SESSION METHOD
    2. CALL TRANSACTION
    3. DIRECT INPUT
    Advantages offered by BATCH INPUT method:
    1. Can process large data volumes in batch.
    2. Can be planned and submitted in the background.
    3. No manual interaction is required when data is transferred.
    4. Data integrity is maintained as whatever data is transferred to the table is through transaction. Hence batch input data is submitted to all the checks and validations.
    To implement one of the supported data transfers, you must often write the program that exports the data from your non-SAP system. This program, known as a “data transfer” program must map the data from the external system into the data structure required by the SAP batch input program.
    The batch input program must build all of the input to execute the SAP transaction.
    Two main steps are required:
    • To build an internal table containing every screen and every field to be filled in during the execution of an SAP transaction.
    • To pass the table to SAP for processing.
    Prerequisite for Data Transfer Program
    Writing a Data Transfer Program involves following prerequisites:
    Analyzing data from local file
    Analyzing transaction
    Analyzing transaction involves following steps:
    • The transaction code, if you do not already know it.
    • Which fields require input i.e., mandatory.
    • Which fields can you allow to default to standard values.
    • The names, types, and lengths of the fields that are used by a transaction.
    • Screen number and Name of module pool program behind a particular transaction.
    To analyze a transaction::
    • Start the transaction by menu or by entering the transaction code in the command box.
    (You can determine the transaction name by choosing System – Status.)
    • Step through the transaction, entering the data will be required for processing your batch input data.
    • On each screen, note the program name and screen (dynpro) number.
    (dynpro = dyn + pro. Dyn = screen, pro = number)
    • Display these by choosing System – Status. The relevant fields are Program (dynpro) and Dynpro number. If pop-up windows occur during execution, you can get the program name and screen number by pressing F1 on any field or button on the screen.
    The technical info pop-up shows not only the field information but also the program and screen.
    • For each field, check box, and radio button on each screen, press F1 (help) and then choose Technical Info.
    Note the following information:
    - The field name for batch input, which you’ll find in its own box.
    - The length and data type of the field. You can display this information by double clicking on the Data Element field.
    • Find out the identification code for each function (button or menu) that you must execute to process the batch-input data (or to go to new screen).
    Place the cursor on the button or menu entry while holding down the left mouse button. Then press F1.
    In the pop-up window that follows, choose Technical info and note the code that is shown in the Function field.
    You can also run any function that is assigned to a function key by way of the function key number. To display the list of available function keys, click on the right mouse button. Note the key number that is assigned to the functions you want to run.
    Once you have program name, screen number, field name (screen field name), you can start writing.
    DATA TRANSFER program.
    Declaring internal table
    First Integral Table similar to structure like local file.
    Declaring internal table like BDCDATA
    The data from internal table is not transferred directly to database table, it has to go through transaction. You need to pass data to particular screen and to particular screen-field. Data is passed to transaction in particular format, hence there is a need for batch input structure.
    The batch input structure stores the data that is to be entered into SAP system and the actions that are necessary to process the data. The batch input structure is used by all of the batch input methods. You can use the same structure for all types of batch input, regardless of whether you are creating a session in the batch input queue or using CALL TRANSACTION.
    This structure is BDCDATA, which can contain the batch input data for only a single run of a transaction. The typical processing loop in a program is as follows:
    • Create a BDCDATA structure
    • Write the structure out to a session or process it with CALL TRANSACTION USING; and then
    • Create a BDCDATA structure for the next transaction that is to be processed.
    Within a BDCDATA structure, organize the data of screens in a transaction. Each screen that is processed in the course of a transaction must be identified with a BDCDATA record. This record uses the Program, Dynpro, and Dynbegin fields of the structure.
    The screen identifier record is followed by a separate BDCDATA record for each value, to be entered into a field. These records use the FNAM and FVAL fields of the BDCDATA structure. Values to be entered in a field can be any of the following:
    • Data that is entered into screen fields.
    • Function codes that are entered into the command field. Such function codes execute functions in a transaction, such as Save or Enter.
    The BDCDATA structure contains the following fields:
    • PROGRAM: Name of module pool program associated with the screen. Set this field only for the first record for the screen.
    • DYNPRO: Screen Number. Set this field only in the first record for the screen.
    • DYNBEGIN: Indicates the first record for the screen. Set this field to X, only for the first record for the screen. (Reset to ‘ ‘ (blank) for all other records.)
    • FNAM: Field Name. The FNAM field is not case-sensitive.
    • FVAL: Value for the field named in FNAM. The FVAL field is case-sensitive. Values assigned to this field are always padded on the right, if they are less than 132 characters. Values must be in character format.
    Transferring data from local file to internal table
    Data is uploaded to internal table by UPLOAD of WS_UPLOAD function.
    Population of BDCDATA
    For each record of internal table, you need to populate Internal table, which is similar to BDCDATA structure.
    All these five initial steps are necessary for any type of BDC interface.
    DATA TRANSFER program can call SESSION METHOD or CALL TRANSACTION. The initial steps for both the methods are same.
    First step for both the methods is to upload the data to internal table. From Internal Table, the data is transferred to database table by two ways i.e., Session method and Call transaction.
    SESSION METHOD
    About Session method
    In this method you transfer data from internal table to database table through sessions.
    In this method, an ABAP/4 program reads the external data that is to be entered in the SAP System and stores the data in session. A session stores the actions that are required to enter your data using normal SAP transaction i.e., Data is transferred to session which in turn transfers data to database table.
    Session is intermediate step between internal table and database table. Data along with its action is stored in session i.e., data for screen fields, to which screen it is passed, the program name behind it, and how the next screen is processed.
    When the program has finished generating the session, you can run the session to execute the SAP transactions in it. You can either explicitly start and monitor a session or have the session run in the background processing system.
    Unless session is processed, the data is not transferred to database table.
    BDC_OPEN_GROUP
    You create the session through program by BDC_OPEN_GROUP function.
    Parameters to this function are:
    • User Name: User name
    • Group: Name of the session
    • Lock Date: The date on which you want to process the session.
    • Keep: This parameter is passed as ‘X’ when you want to retain session after
    processing it or ‘ ‘ to delete it after processing.
    BDC_INSERT
    This function creates the session & data is transferred to Session.
    Parameters to this function are:
    • Tcode: Transaction Name
    • Dynprotab: BDC Data
    BDC_CLOSE_GROUP
    This function closes the BDC Group. No Parameters.
    Some additional information for session processing
    When the session is generated using the KEEP option within the BDC_OPEN_GROUP, the system always keeps the sessions in the queue, whether it has been processed successfully or not.
    However, if the session is processed, you have to delete it manually. When session processing is completed successfully while KEEP option was not set, it will be removed automatically from the session queue. Log is not removed for that session.
    If the batch-input session is terminated with errors, then it appears in the list of INCORRECT session and it can be processed again. To correct incorrect session, you can analyze the session. The Analysis function allows to determine which screen and value has produced the error. If you find small errors in data, you can correct them interactively, otherwise you need to modify batch input program, which has generated the session or many times even the data file.
    CALL TRANSACTION
    About CALL TRANSACTION
    A technique similar to SESSION method, while batch input is a two-step procedure, Call Transaction does both steps online, one after the other. In this method, you call a transaction from your program by
    Call transaction <tcode> using <BDCTAB>
    Mode <A/N/E>
    Update <S/A>
    Messages into <MSGTAB>.
    Parameter – 1 is transaction code.
    Parameter – 2 is name of BDCTAB table.
    Parameter – 3 here you are specifying mode in which you execute transaction
    A is all screen mode. All the screen of transaction are displayed.
    N is no screen mode. No screen is displayed when you execute the transaction.
    E is error screen. Only those screens are displayed wherein you have error record.
    Parameter – 4 here you are specifying update type by which database table is updated.
    S is for Synchronous update in which if you change data of one table then all the related Tables gets updated. And sy-subrc is returned i.e., sy-subrc is returned for once and all.
    A is for Asynchronous update. When you change data of one table, the sy-subrc is returned. And then updating of other affected tables takes place. So if system fails to update other tables, still sy-subrc returned is 0 (i.e., when first table gets updated).
    Parameter – 5 when you update database table, operation is either successful or unsuccessful or operation is successful with some warning. These messages are stored in internal table, which you specify along with MESSAGE statement. This internal table should be declared like BDCMSGCOLL, a structure available in ABAP/4. It contains the following fields:
    1. Tcode: Transaction code
    2. Dyname: Batch point module name
    3. Dynumb: Batch input Dyn number
    4. Msgtyp: Batch input message type (A/E/W/I/S)
    5. Msgspra: Batch input Lang, id of message
    6. Msgid: Message id
    7. MsgvN: Message variables (N = 1 - 4)
    For each entry, which is updated in database, table message is available in BDCMSGCOLL. As BDCMSGCOLL is structure, you need to declare a internal table which can contain multiple records (unlike structure).
    Steps for CALL TRANSACTION method
    1. Internal table for the data (structure similar to your local file)
    2. BDCTAB like BDCDATA
    3. UPLOAD or WS_UPLOAD function to upload the data from local file to itab. (Considering file is local file)
    4. Loop at itab.
    Populate BDCTAB table.
    Call transaction <tcode> using <BDCTAB>
    Mode <A/N/E>
    Update <S/A>.
    Refresh BDCTAB.
    Endloop.
    (To populate BDCTAB, You need to transfer each and every field)
    The major differences between Session method and Call transaction are as follows:
    SESSION METHOD CALL TRANSACTION
    1. Data is not updated in database table unless Session is processed. Immediate updation in database table.
    2. No sy-subrc is returned. Sy-subrc is returned.
    3. Error log is created for error records. Errors need to be handled explicitly
    4. Updation in database table is always synchronous Updation in database table can be synchronous Or Asynchronous.
    Error Handling in CALL TRANSACTION
    When Session Method updates the records in database table, error records are stored in the log file. In Call transaction there is no such log file available and error record is lost unless handled. Usually you need to give report of all the error records i.e., records which are not inserted or updated in the database table. This can be done by the following method:
    Steps for the error handling in CALL TRANSACTION
    1. Internal table for the data (structure similar to your local file)
    2. BDCTAB like BDCDATA
    3. Internal table BDCMSG like BDCMSGCOLL
    4. Internal table similar to Ist internal table
    (Third and fourth steps are for error handling)
    5. UPLOAD or WS_UPLOAD function to upload the data from the local file to itab. (Considering file is local file)
    6. Loop at itab.
    Populate BDCTAB table.
    Call transaction <tr.code> using <Bdctab>
    Mode <A/N/E>
    Update <S/A>
    Messages <BDCMSG>.
    Perform check.
    Refresh BDCTAB.
    Endloop.
    7 Form check.
    IF sy-subrc <> 0. (Call transaction returns the sy-subrc if updating is not successful).
    Call function Format_message.
    (This function is called to store the message given by system and to display it along with record)
    Append itab2.
    Display the record and message.
    DIRECT INPUT
    About Direct Input
    In contrast to batch input, this technique does not create sessions, but stores the data directly. It does not simulate the online transaction. To enter the data into the corresponding database tables directly, the system calls a number of function modules that execute any necessary checks. In case of errors, the direct input technique provides a restart mechanism. However, to be able to activate the restart mechanism, direct input programs must be executed in the background only. Direct input checks the data thoroughly and then updates the database directly.
    You can start a Direct Input program in two ways;
    Start the program directly
    This is the quickest way to see if the program works with your flat file. This option is possible with all direct input programs. If the program ends abnormally, you will not have any logs telling you what has or has not been posted. To minimize the chance of this happening, always use the check file option for the first run with your flat file. This allows you to detect format errors before transfer.
    Starting the program via the DI administration transaction
    This transaction restarts the processing, if the data transfer program aborts. Since DI document are immediately posted into the SAP D/B, the restart option prevents the duplicate document posting that occurs during a program restart (i.e., without adjusting your flat file).
    Direct input is usually done for standard data like material master, FI accounting document, SD sales order and Classification for which SAP has provided standard programs.
    First time you work with the Direct Input administration program, you will need to do some preparation before you can transfer data:
    - Create variant
    - Define job
    - Start job
    - Restart job
    Common batch input errors
    - The batch input BDCDATA structure tries to assign values to fields which do not exist in the current transaction screen.
    - The screen in the BDCDATA structure does not match the right sequence, or an intermediate screen is missing.
    - On exceptional occasions, the logic flow of batch input session does not exactly match that of manual online processing. Testing the sessions online can discover by this.
    - The BDCDATA structure contains fields, which are longer than the actual definition.
    - Authorization problems.
    RECORDING A BATCH INPUT
    A B recording allows you to record a R/3 transaction and generate a program that contains all screens and field information in the required BDC-DATA format.
    You can either use SHDB transaction for recording or
    SYSTEM ? SERVICES ? BATCH INPUT ? EDIT
    And from here click recording.
    Enter name for the recording.
    (Dates are optional)
    Click recording.
    Enter transaction code.
    Enter.
    Click Save button.
    You finally come to a screen where, you have all the information for each screen including BDC_OKCODE.
    • Click Get Transaction.
    • Return to BI.
    • Click overview.
    • Position the cursor on the just recorded entry and click generate program.
    • Enter program name.
    • Click enter
    The program is generated for the particular transaction.
    BACKGROUND PROCESSING
    Need for Background processing
    When a large volume of data is involved, usually all batch inputs are done in background.
    The R/3 system includes functions that allow users to work non-interactively or offline. The background processing systems handle these functions.
    Non-interactively means that instead of executing the ABAP/4 programs and waiting for an answer, user can submit those programs for execution at a more convenient planned time.
    There are several reasons to submit programs for background execution.
    • The maximum time allowed for online execution should not exceed 300 seconds. User gets TIMEOUT error and an aborted transaction, if time for execution exceeds 300 seconds. To avoid these types of error, you can submit jobs for background processing.
    • You can use the system while your program is executing.
    This does not mean that interactive or online work is not useful. Both type of processing have their own purposes. Online work is the most common one entering business data, displaying information, printing small reports, managing the system and so on. Background jobs are mainly used for the following tasks; to process large amount of data, to execute periodic jobs without human intervention, to run program at a more convenient, planned time other than during normal working hours i.e., Nights or weekends.
    The transaction for background processing is SM36.
    Or
    Tools ? Administration ? Jobs ? Define jobs
    Or
    System ? services ? Jobs
    Components of the background jobs
    A job in Background processing is a series of steps that can be scheduled and step is a program for background processing.
    • Job name. Define the name of assigned to the job. It identifies the job. You can specify up to 32 characters for the name.
    • Job class. Indicates the type of background processing priority assigned to the job.
    The job class determines the priority of a job. The background system admits three types of job classes: A B & C, which correspond to job priority.
    • Job steps. Parameters to be passed for this screen are as follows:
    Program name.
    Variant if it is report program
    Start criteria for the job: Option available for this are as follows:
    Immediate - allows you to start a job immediately.
    Date/Time - allows you to start a job at a specific name.
    After job - you can start a job after a particular job.
    After event - allows you to start a job after a particular event.
    At operation mode - allows you to start a job when the system switches to a particular operation mode.
    Defining Background jobs
    It is two step process: Firstly, you define the job and then release it.
    When users define a job and save it, they are actually scheduling the report i.e., specifying the job components, the steps, the start time.
    When users schedule program for background processing, they are instructing the system to execute an ABAP/4 report or an external program in the background. Scheduled jobs are not executed until they are released. When jobs are released, they are sent for execution to the background processing system at the specified start time. Both scheduling and releasing of jobs require authorizations.
    HANDLING OF POP UP SCREEN IN BDC
    Many times in transaction pop up screen appears and for this screen you don’t pass any record but some indication to system telling it to proceed further. For example: The following screen
    To handle such screen, system has provided a variable called BDC_CURSOR. You pass this variable to BDCDATA and process the screen.
    Usually such screen appears in many transactions, in this case you are just passing information, that YES you want to save the information, that means YES should be clicked. So you are transferring this information to BDCDATA i.e., field name of YES which is usually SPOT_OPTION. Instead of BDC_OKCODE, you are passing BDC_CURSOR.
    BDC_CURSOR is also used to place cursor on particular field.
    A simple transaction where you are entering customer number on first screen and on next screen data is displayed for the particular customer number. Field, which we are changing here, are name and city. When you click on save, the changed record gets saved.
    Prerequisite to write this BDC interface as indicated earlier is:
    1. To find screen number
    2. To find screen field names, type of the field and length of the field.
    3. To find BDC_OKCODE for each screen
    4. Create flat file.
    generally  Batch Input usually are used to transfer large amount of data. For example you are implementing a new SAP project, and of course you will need some data transfer from legacy system to SAP system.
    CALL TRANSACTION is used especially for integration actions between two SAP systems or between different modules. Users sometimes wish to do something like that click a button or an item then SAP would inserts or changes data automatically. Here CALL TRANSACTION should be considered.
    2. Transfer data for multiple transactions usually the Batch Input method is used.
    check these sites for step by step process:
    For BDC:
    http://myweb.dal.ca/hchinni/sap/bdc_home.htm
    https://www.sdn.sap.com/irj/sdn/wiki?path=/display/home/bdc&
    http://www.sap-img.com/abap/learning-bdc-programming.htm
    http://www.sapdevelopment.co.uk/bdc/bdchome.htm
    http://www.sap-img.com/abap/difference-between-batch-input-and-call-transaction-in-bdc.htm
    http://help.sap.com/saphelp_47x200/helpdata/en/69/c250684ba111d189750000e8322d00/frameset.htm
    http://www.sapbrain.com/TUTORIALS/TECHNICAL/BDC_tutorial.html
    Check these link:
    http://www.sap-img.com/abap/difference-between-batch-input-and-call-transaction-in-bdc.htm
    http://www.sap-img.com/abap/question-about-bdc-program.htm
    http://www.itcserver.com/blog/2006/06/30/batch-input-vs-call-transaction/
    http://www.planetsap.com/bdc_main_page.htm
    call Transaction or session method ?
    null

  • Caching objects in the data cache as a result of an extent.

    Patrick -
    I wanted to post this since it's related to a question I posted about extents and the data cache on
    11/8.
    I discovered that the com.solarmetric.kodo.DefaultFetchBatchSize setting affects how many objects
    get put into the data cache as a result of running an extent (in 2.3.2). If I have:
    com.solarmetric.kodo.DefaultFetchBatchSize=20
    then as soon as I execute the second line below:
    Iterator anIterator = results.iterator();
    Object anObject = anIterator.next();
    I see 20 objects in my data cache. In a prior reply you indicated that you were going to check this
    behavior in 2.4 so I wanted to send you this additional information. This behavior isn't a problem
    for me.
    Les

    Les,
    This is expected behavior -- the DefaultBatchFetchSize instructs Kodo to
    retrieve objects from the scrollable ResultSet in groups of 20. So,
    getting the first item from the iterator will cause a page of 20 objects
    to be pulled from the result set.
    -Patrick
    Les Selecky wrote:
    Patrick -
    I wanted to post this since it's related to a question I posted about
    extents and the data cache on
    11/8.
    I discovered that the com.solarmetric.kodo.DefaultFetchBatchSize
    setting affects how many objects
    get put into the data cache as a result of running an extent (in
    2.3.2). If I have:
    com.solarmetric.kodo.DefaultFetchBatchSize=20
    then as soon as I execute the second line below:
    Iterator anIterator = results.iterator();
    Object anObject = anIterator.next();
    I see 20 objects in my data cache. In a prior reply you indicated that
    you were going to check this
    behavior in 2.4 so I wanted to send you this additional information.
    This behavior isn't a problem
    for me.
    Les
    Patrick Linskey [email protected]
    SolarMetric Inc. http://www.solarmetric.com

  • Can "trigger" management be addressed in the data modeler?

    Please add an easier way to add/modify triggers.
    I reverse-engineered my model into the data modeler and my triggers all migrated correctly to the model but, for the life of me, had no idea how to view, modify or create triggers.  After extensive research, I found that they can be accessed in the left-hand Browser by drilling some 8 levels to view the triggers for each table - why?  Why can't this functionality be added to properties context menu for the table.  I'm not a big advocate for triggers but if I need to work on one, the current methodology is tedious at best.
    Of course, if there's some other way to do this, please inform...
    Accessing Triggers via the Browser
    Design -> Relational Model -> APEX_DEMO -> Physical Model -> Oracle Database 11g -> Tables -> ACTIVITY_LOG -> Triggers

    Thanks Joop and Phillip...
    While I occasionally use SQL Developer, I currently use PL/SQL Developer (Allround Automations) as my development tool of choice and bringing a 3rd tool as SQL Developer into the mix, makes no sense to me.
    I completely agree that managing a trigger via a development tool is much easier but one occasionally wants to quickly add a trigger either manually or via a template and this is not easy currently.
    While functions, procedures, package, etc. are all important to a data model, the existence of most triggers are completely dependent on a table and are a critical component in any model that uses them.  The ability to access and modify them easily in Data Modeler is important and shouldn't be an afterthought as is perceived here.  If trigger access cannot be added to the Table Properties (as noted in the image below), at the very least, please provide a shortcut methodology to do so.

  • Inserting rows into a data block

    Hi, how do i insert 'n' rows into a data block and after that commit the changes into the data base.
    Thanks for any help.

    You can use something like this:
    Go_Block('my_block');
    first_record;
    LOOP
    .... insert values to the items
    next_record;
    END LOOP;
    - for commiting changes
    Do_Key('Commit_Form');
    I hope this will help you.
    Helena

  • To pull data from avarchar field into a date field in another table

    Hi
    I pulled some records into a table from a remote server by means of a database link and one of the field which is varchar(30) needs to be pulled into other database which has a date column.
    so the varchar field needs to be converted into the date column so that the data can be loaded into that table.

    Now, we are talking. See, all you had to give us was the error first. Anyway, I think your 'format' coming in is different from the date 'format' you are using to do the 'to_date' function.
    Anyway, please show us what the incoming data looks like and we might be able to suggest what is the 'to_date' format you have to use (for eg: 'yyyy-mm-dd' or 'yyyy mm dd HH:mi:ssss').

Maybe you are looking for

  • Kontakt 4.1.1 and Logic 9.1.1 needs Max buffer

    I posted this on the NI forum too.I have 1 instance of Kontakt 4.1.1 in Logic 9.1.1 64bit with only 4 sounds loaded and have to set the buffer to Max 1024 otherwise all I get is clicks and pops on playback.The strange part is,when I first loaded this

  • Program ZRFFOD__S: No records selected

    Hi I run F110 then the documents has been cleared but when i checked spool request I cant see any. I checked it in payment log and find this error "Program ZRFFOD__S: No records selected". Is there anyone know why this error has been encountered? tha

  • Treo800w does not use the ringer volume I have set

    I have my 'Phone: Known Caller' ringer setting set to the standard Treo ring tone that came preinstalled on the Treo800w. I have the 'Ring type' set to Ring but NO matter how high or low I set the ringer volume it only rings at the lowest volume sett

  • Invoice issue -transfer to FI dat

    hii all, please resolve this issue with clarification is point-by-point....... "Invoice amount is completely correct, but if you look at the trip details in the body of invoice they do not match. On the service billing document the expense detail on

  • Installing XMLSEC in Mac Os X without installing xCode

    I am an Applescript programmer from Brazil. I have developed a solution that uses xmlsec to sign an xml file that contains an invoice. (this is a new procedure in my country, all invoices have to be signed by the companies and aproved by government b