Mass data+use bulk itu00B4s possible ?

hi guys,
I have in my backend system about 200.000 register.
I use bulk (get_detail) to bring all data using a single get_detail , instead of use get_list and after get the keys headers execute details for each key (200000 times) .
The approach is correct ? what the best solution ? can I use mass + use bulk ?
I need help about huge data transfer to CDS. thanks

You can use Mass GetList and Bulk Get Detail together.
If you are working with SP04 of Gateway1.1 to SAP Netweaver Mobile 7.10. Then you can check the code in following data objcets.
1. Goto SDOE_WB transaction
2. Select MAS_SRVASG data object from SWCV :: MAS_SMARTPHONE_SERVICE , 1.4 of  SAP.
3. Select the backend adapter "MAS_SRVASG_BA" and view the Get List & Get Details BAPI wrappers which uses both these functionalities.
In Short:
Get List with Mass Data: This should provide set of record keys. [Using Mass Get List|http://www.sdn.sap.com/irj/scn/index?rid=/library/uuid/50fa3a06-8bcb-2c10-1daf-c604a4880361]
Get Detail with Bulk: This should have root node details as TABLES parameter. DOE will invoke the BAPI wrapper by providing number of keys in this TABLES parameter. BAPI wrapper should return the values for all these keys by populating all other tables and the root node information also. Check question 8 from following link:
[Question about Use Bulk Detail|http://www.sdn.sap.com/irj/scn/index?rid=/library/uuid/10de86e2-41d0-2c10-7e96-e9337967f7fe]
Regards,
Siva.
Edited by: Siva Satya Prasad Yerra on May 9, 2011 12:44 PM

Similar Messages

  • Best option to push Mass data using Java API

    Hi All,
    Can any one let me know how Java API push data to MDM? Is it a good option when ti comes to perfromance point of view? We have around 2000/5000 records to be imported every time using Java API and wondering what would be best way to acheive this using Java API?
    Is java API is best option than Web services?
    Thanks
    Rajeev

    Rajeev,
    It is perfectly fine to use API to push data into MDM, you just need to make sure that your API does all those tasks which import manager does by default. For example, it should know when to insert and when to update. if an error happens during insertion of record, you should be able to notify the users.
    Having said that, where does the number 2000/5000 comes in? will you be creating users in batch and then invoke MDM API to store these in a single shot or willl it be a delta insert/update from Java to MDM?
    Thanks
    Aamir

  • Parallel processing of mass data : sy-subrc value is not changed

    Hi,
    I have used the Parallel processing of mass data using the "Start New Task" . In my function module I am handling the exceptions and finally raise the application specific old exception to be handled in my main report program. Somehow the sy-subrc is not getting changed and always returns 0 even if the expection is raised.
    Can anyone help me about the same.
    Thanks & Regards,
    Nitin

    Hi Silky,
    I've build a block of code to explain this.
      DATA: ls_edgar TYPE zedgar,
            l_task(40).
      DELETE FROM zedgar.
      COMMIT WORK.
      l_task = 'task1'.
      ls_edgar-matnr = '123'.
      ls_edgar-text = 'qwe'.
      CALL FUNCTION 'Z_EDGAR_COMMIT_ROLLBACK' STARTING NEW TASK l_task PERFORMING f_go ON END OF TASK
        EXPORTING
          line = ls_edgar.
      l_task = 'task2'.
      ls_edgar-matnr = 'abc'.
      ls_edgar-text = 'def'.
      CALL FUNCTION 'Z_EDGAR_COMMIT_ROLLBACK' STARTING NEW TASK l_task PERFORMING f_go ON END OF TASK
        EXPORTING
          line = ls_edgar.
      l_task = 'task3'.
      ls_edgar-matnr = '456'.
      ls_edgar-text = 'xyz'.
      CALL FUNCTION 'Z_EDGAR_COMMIT_ROLLBACK' STARTING NEW TASK l_task PERFORMING f_go ON END OF TASK
        EXPORTING
          line = ls_edgar.
    *&      Form  f_go
    FORM f_go USING p_c TYPE ctype.
      RECEIVE RESULTS FROM FUNCTION 'Z_EDGAR_COMMIT_ROLLBACK' EXCEPTIONS err = 2.
      IF sy-subrc = 2.
    *this won't affect the LUW of the received function
        ROLLBACK WORK.
      ELSE.
    *this won't affect the LUW of the received function
        COMMIT WORK.
      ENDIF.
    ENDFORM.                    "f_go
    and the function is:
    FUNCTION z_edgar_commit_rollback.
    *"*"Interface local:
    *"  IMPORTING
    *"     VALUE(LINE) TYPE  ZEDGAR
    *"  EXCEPTIONS
    *"      ERR
      MODIFY zedgar FROM line.
      IF line-matnr CP 'a*'.
    *comment raise or rollback/commit to test
    *    RAISE err.
        ROLLBACK WORK.
      ELSE.
        COMMIT WORK.
      ENDIF.
    ENDFUNCTION.
    ok.
    In your main program you have a Logical Unit of Work (LUW), witch consists of an application transaction and is associated with a database transaction. Once you start a new task, your creating an independent LUW, with it's own database transaction.
    So if you do a commit or rollback in your function the effect is only on the records your processing in the function.
    There is a way to capture the event when this LUW concludes in the main LUW. That is the PERFORMING whatever ON END OF TASK. In there you can get the result of the function but you cannot commit or rollback the LUW from the function since it already have implicitly happened at the conclusion of the funtion. You can test it by correctly comment the code I've supplied.
    So, if you  want to rollback the LUW of the function you better do it inside it.
    I don't think it matches exactly your question, maybe it lead you on the right track. Give me more details if it doesn't.
    Hope it helps,
    Edgar

  • Error using BULK Collect with RECORD TYPE

    hello
    I have written a simple Procedure by declaring a record type & then making a variable of NESTED Table type.
    I then select data using BULK COLLECT & tryin to access it through a LOOP.....Getting an ERROR.
    CREATE OR REPLACE PROCEDURE sp_test_bulkcollect
    IS
    TYPE rec_type IS RECORD (
    emp_id VARCHAR2(20),
    level_id NUMBER
    TYPE v_rec_type IS TABLE OF rec_type;
    BEGIN
    SELECT employee_id, level_id
    BULK COLLECT INTO v_rec_type
    FROM portfolio_exec_level_mapping
    WHERE portfolio_execp_id = 2851852;
    FOR indx IN v_rec_type.FIRST..v_rec_type.LAST
    LOOP
    dbms_output.put_line('Emp -- '||v_rec_type.emp_id(indx)||' '||v_rec_type.level_id(indx));
    END LOOP;
    END;
    Below are the ERROR's i am getting ....
    - Compilation errors for PROCEDURE DOMRATBDTESTUSER.SP_TEST_BULKCOLLECT
    Error: PLS-00321: expression 'V_REC_TYPE' is inappropriate as the left hand side of an assignment statement
    Line: 15
    Text: FROM portfolio_exec_level_mapping
    Error: PL/SQL: ORA-00904: : invalid identifier
    Line: 16
    Text: WHERE portfolio_execp_id = 2851852;
    Error: PL/SQL: SQL Statement ignored
    Line: 14
    Text: BULK COLLECT INTO v_rec_type
    Error: PLS-00302: component 'FIRST' must be declared
    Line: 19
    Text: LOOP
    Error: PL/SQL: Statement ignored
    Line: 19
    Text: LOOP
    PLZ Help.

    and with a full code sample:
    SQL> CREATE OR REPLACE PROCEDURE sp_test_bulkcollect
      2  IS
      3  TYPE rec_type IS RECORD (
      4  emp_id VARCHAR2(20),
      5  level_id NUMBER
      6  );
      7  TYPE v_rec_type IS TABLE OF rec_type;
      8  v v_rec_type;
      9  BEGIN
    10     SELECT empno, sal
    11     BULK COLLECT INTO v
    12     FROM emp
    13     WHERE empno = 7876;
    14     FOR indx IN v.FIRST..v.LAST
    15     LOOP
    16        dbms_output.put_line('Emp -- '||v(indx).emp_id||' '||v(indx).level_id);
    17     END LOOP;
    18  END;
    19  /
    Procedure created.
    SQL>
    SQL> show error
    No errors.
    SQL>
    SQL> begin
      2     sp_test_bulkcollect;
      3  end;
      4  /
    Emp -- 7876 1100
    PL/SQL procedure successfully completed.

  • Is it possible to read digital data using an external clock (PCI-6259 M)?

    I’m using a NI PCI-6259 M Series card and trying to write my program in VC++6.0 using the functions in the DAQmx driver.
    Question1: Not all functions listed in the NI-DAQmx C Reference Help seems to be supported by my NI-card, where can I find information about which of the functions that are supported?
    Question2: I want to read data from a device that clock out data on the falling edge of a clock signal. The clock signal and the data signal are routed to two DIO terminals on the NI-card. The question is if it is possible to read data using the clock as a sample clock? See two code examples below that doesn’t work. In both cases 10 samples are read at once, even if the external clock is not present.
    Example 1
    // Create tasks
    Status = DAQmxCreateTask("", &m_ReadTrimTask);
    // Set up read task
    status = DAQmxCreateDIChan(m_ReadTrimTask, "Dev1/port2/line0", "", DAQmx_Val_ChanPerLine);
    status = DAQmxCfgChangeDetectionTiming(m_ReadTrimTask,"Dev1/port2/line6","Dev1/port2/line6",DAQmx_Val_FiniteSamps, 10);
    // Read data
    int32 sampsPerChanRead, numBytesPerSamp;
    status = DAQmxReadDigitalLines(m_ReadTrimTask, 10, 10.0, DAQmx_Val_GroupByChannel, result, 10, &sampsPerChanRead, &numBytesPerSamp ,NULL);
    Example 2
    // Create tasks
    Status = DAQmxCreateTask("", &m_ReadTrimTask);
    // Set up read task
    status = DAQmxCreateDIChan(m_ReadTrimTask, "Dev1/port2/line0", "", DAQmx_Val_ChanPerLine);
    status = DAQmxSetSampTimingType(m_ReadTrimTask, DAQmx_Val_SampClk);
    status = DAQmxSetSampClkRate(m_ReadTrimTask, 1000.0);
    status = DAQmxSetSampClkActiveEdge(m_ReadTrimTask, DAQmx_Val_Falling);
    status = DAQmxSetSampClkSrc(m_ReadTrimTask, " Dev1/port2/line6");
    // Read data
    int32 sampsPerChanRead, numBytesPerSamp;
    status = DAQmxReadDigitalLines(m_ReadTrimTask, 10, 10.0, DAQmx_Val_GroupByChannel, result, 10, &sampsPerChanRead, &numBytesPerSamp ,NULL);

    Hello Magnus,
    Thank you for contacting National Instruments.
    "Question1: Not all functions listed in the NI-DAQmx C Reference Help seems to be supported by my NI-card, where can I find information about which of the functions that are supported?"
    The best place to look for this information would be the M Series Help Manual. There you can find the features of your PCI-6259 and what operations it supports.
    "Question2: I want to read data from a device that clock out data on the falling edge of a clock signal. The clock signal and the data signal are routed to two DIO terminals on the NI-card. The question is if it is possible to read data using the clock as a sample clock? See two code examples below that doesn’t work. In both cases 10 samples are read at once, even if the external clock is not present."
    Look at the "ContReadDigChan-ExtClk_Fn.c" example project which ships with the NI-DAQ driver. This is located at: C:\Program Files\National Instruments\NI-DAQ\Examples\DAQmx ANSI C\Digital\Read Values\Cont Read Dig Chan-Ext Clk.
    You will have to make some minor modifications to convert this to a finite acquisition, but that is simply a matter of changing the "sampleMode" parameter of the DAQmxCfgSampClkTiming() function. You will also have to route your clock signal to a PFI line and specify which line in your code.
    I hope this helps.
    Sean C.
    Applications Engineering
    National Instruments

  • Problem in Using Mass Data Getlist.

    Hi All,
    I am using a "Mass Data" Getlist for Salesorder. there are nearly 8000 records in VBAP table wich I want to bring it in Batches of 1000.
    I have modified my Getlist to the requirement as given in the document
    [http://www.sdn.sap.com/irj/scn/go/portal/prtroot/docs/library/uuid/50fa3a06-8bcb-2c10-1daf-c604a4880361?quicklink=index&overridelayout=true]
    FUNCTION Z_SALES_MASS_GETLIST.
    *"*"Local Interface:
    *"  IMPORTING
    *"     VALUE(ME_MASS_REP) LIKE  MEREP_MASS_REP1 STRUCTURE
    *"        MEREP_MASS_REP1 OPTIONAL
    *"  EXPORTING
    *"     VALUE(RETURN) TYPE  BAPIRET2
    *"  TABLES
    *"      IT_ZSALESORDERHEADER STRUCTURE  ZSALESORDERHEADER
    DATA : LDF_VBELN LIKE VBAK-VBELN.
    LDF_VBELN = ME_MASS_REP-ME_KEY_GREATER_THAN.
    select *
    from   VBAK
    INTO   CORRESPONDING FIELDS OF TABLE IT_ZSALESORDERHEADER
      UP TO     ME_MASS_REP-me_number_of_headers ROWS
                       WHERE     VBELN > LDF_VBELN
                       ORDER BY PRIMARY KEY.
    ENDFUNCTION.
    The Problem is that when I call this BAPI  in the DOE using the adapter and trigger the initial download I am not getting in any data into the middlware from the backend.
    Please let me know where do I pass the values of ME_KEY_GREATER_THAN and ME_NUMBER_OF_HEADERS.
    Is it in the backend or through the middleware?
    Is that the reason why I am not getting in the data to the middleware?
    Edited by: krutheeka on May 3, 2011 6:31 AM
    Edited by: krutheeka on May 3, 2011 6:32 AM

    Hi ,
    I have placed a breakpoint as I mentioned previously in the Generated class GET_KEY at the code,
    e_IT_ZSALESORDERHEADER = lt_header.
    I am getting  the data into e_IT_SALESORDERHEADER. I have mentioned that I need data in batches of 1000 so when I
    excecute this method I get the records into the above structure in batch of 1000.
    but I am not able to see any data in the CDS.
    Someone please tell me what the problem might be..
    Thanks,
    Krutheeka.

  • If your database in Full Recovery mode, can you use Bulk Insert Task to load data

    If your database in Full Recovery mode, can you use Bulk Insert Task to load data

    If your database in Full Recovery mode, can you use Bulk Insert Task to load data
    Yes you can ofourse but dont be in idea that logging will be mininal. Loggign will be as per recovery model full. Every thing will be logged. If you are going to use bulk insert task you can consider switching recovery model to Bulk logged but you will not
    have option to do point in time recovery.
    PS: please dont create duplicate threads
    If you read first Note section in below link it clearly states that yes logging will be full and you can use
    http://technet.microsoft.com/en-us/library/ms191244(v=sql.105).aspx
    Please mark this reply as answer if it solved your issue or vote as helpful if it helped so that other forum members can benefit from it.
    My TechNet Wiki Articles

  • Is it possible to retrieve data using RFC alone with out using BAPI ?

    Hi ,
    Is it possible to connect SAP to Non-SAP systems to retrieve data using RFC alone with out using BAPI ?
    thanks
    Rama

    Hello Nelson,
    Could you please explain how this can be done. Because my understanding is that only BAPI can be used to connect Non SAP and SAP system not RFC function module.
    Regards
    Farzan

  • HT4203 I have turned off the Data use on the I-phone 4s but am told by AT&T that data is still being used for applications.  How is this possible. Can someone help me understand this.  Thanks.

    I have turned off the Data use on the I-phone 4s but am told by AT&T that data is still being used for applications.  How is this possible. Can someone help me understand this. 

    I assume you are including the STIX font as part of your epub files?     
    Perhaps the folks who do this blog might be able to help -- they have done some work with font embedding:
    http://www.pigsgourdsandwikis.com/2011/04/embedding-fonts-in-epub-ipad-iphone-an d.html

  • I have a new macbook air, and want to transfer all my data using time machine from my old macbook running leopard, is this possible and easy

    I have a new macbook air, and want to transfer all my data using time machine from my old macbook running leopard, is this possible and easy and will it transfer all my musc and photos?

    The Setup Assistant can restore your data from a Time Machine backup. You'll be asked if you want to do this the first time you turn the computer on.
    (60797)

  • Developing SAP Reports Using Mass Data Runtime( MDR)?

    Hi Experts,
    Does anyone have the idea of MDR reporting?
    Actually i have implemented this in my previous project which is used for parallel processing.
    The problem which I am facing currently is that when i try to open the transaction /BTR/MDR it is throwing an error as "transaction doesn't exits'.
    Why this transaction is not opening in the system? Please guide me on this issue
    Thanks in Advance

    Hi Nawaz,
    Mass Data Runtime (MDR) is a third party software component. If it has not been purchased and installed in the system you are working on, it is unlikely that the transaction will exist.
    Hope that helps.
    Christian

  • I had iphone 5s which was accidently dropped in water i dont know if it could be fixed or not but is it possible to save my data using i clouds without fixing the phone

    i had iphone 5s which was accidently dropped in water i dont know if it could be fixed or not but is it possible to save my data using i clouds without fixing the phone

    if you set it up to sync with iclouds in terms of contacts and calendar and the likes in advanced the data is still in icloud
    if you set it up to backup to iclouds since you only get 5GB free storage you would likely had paid to get more or used very little storage on your phone then the backup is still in iclouds
    about if it can be fixed then read
    iPhone 5 water - how screwed am i

  • Is there a way to dynamically figure out the FIRSTROW & LASTROW using Bulk Insert?

    I noticed most of my files follow a certain convention, so I can use Bulk Insert with this:
    FIRSTROW = 2, LASTROW = 224
    However a few files have many more records.  I'd like SQL Server to dynamically figure out the first and last rows with data, but I don't know if that's possible.  Can someone confirm?
    Knowledge is the only thing that I can give you, and still retain, and we are both better off for it.

    Nope, you would need read the files from some program. Could be a CLR stored procedure if you want to do it from SQL Server. But from BULK INSERT alone, no.
    Erland Sommarskog, SQL Server MVP, [email protected]

  • How to use bulk insert in utl_file

    i am creating text file of employee list i go huge data i want to use bulk insert with utl_file is it possible if so, how pls help me.

    If what you are asking, in some unknown version of Oracle, is whether using FORALL will increase the speed for writing a file to your hard disk the answer is no.
    BULK COLLECT, however, might help.

  • Mass data update in Value mapping table

    Hi ,
      I have used Value mapping replication to update the mass data from external source to Value mapping table . Its updatating in Runtime Cache but i want the data to be visible in GUI Value mapping table as well. Is it possible ? Because i doubt that the data in Runtime Cache may get  removed if the system restarts . Can any one help ?
    Thanks
    Laks

    Hi NALLAM GUNA RANJAN,
      Thanks for your prompt reply but i didn't get what you are trying to convey . My issue here is
    Instead of manually entering key-value pair in Value mapping table , I used Value mapping replication ( http://help.sap.com/saphelp_nw04/helpdata/en/2a/9d2891cc976549a9ad9f81e9b8db25/frameset.htm )
    Its updating the data in Runtime Cache ( you can see this using Cache Monitoring )  but not able to view the data in Actual Value mapping Table (GUI in Directory of SAPXI) I want the data updated using Replication to be visible in GUI table is it possible ?
    Hope you got the question much better now
    Thanks
    Laks

Maybe you are looking for

  • My BB Pearl's 9105 trackpad won't work!!!!

    Somebody help me out on this! My BB Pearl's trackpad is not working! It won't scroll and select neither! Gosh! What should i do about this? Help me please! Pretty please!

  • Widescreen Problems

    Hi, Hopefully this is in the right place. I have transferred some recordings from my digibox to my Mac, which start life as a ts format. I've converted these using iSquint to mpeg-4. I can view the new file absolutely fine on VLC Media Player in the

  • XK01 Uploading thru LSMW

    while uploading Vendor master thru LSMW, if we will create one record to get the field name when i tried that the screen is looking different in the template which i have collected from client is having 2 columns for address. Street 2, street 3 Now i

  • Screen Sharing: reducing network overhead?

    After setting up screen sharing, are there some ways of keeping the overhead low in terms of CPU and hard drive load? I noticed after getting screen sharing going, a fair amount of hard drive chatter on the computer who's screen was being shared; not

  • When trying to install CC, I get error A12E1.

    I have tried to intall CC both before and after upgrading to OSX Mavericks.  I still get error A12E1. System info: Hardware Overview:   Model Name:          MacBook Pro   Model Identifier:          MacBookPro8,1   Processor Name:          Intel Core