Open Hub Destinations - Write directly to a table on 3rd party DB.

Hi All,
Is there way to create a table in the Non SAP system ( like 3rd party DB ) using Open Hub Destinations?
If yes what are the steps.
Thanks,
Hima

There is a 3rd party extractor to load data from BI into SQL Server (and vice versa) using Open Hub:
[ http://msdn.microsoft.com/en-us/library/dd299430.aspx|http://msdn.microsoft.com/en-us/library/dd299430.aspx]
Summary: This white paper demonstrates the use of the Microsoft Connector 1.0 for SAP BI in Microsoft SQL Server 2008 Integration Services packages. It shows how to load data into SAP BI by using the SAP BI destination, how to extract data from SAP BI by using the SAP BI source, and how to prepare extracted data for analysis in SQL Server Analysis Services.

Similar Messages

  • List of Open Hub Destination

    Hi All,
    Is there a table which list out all the open hub destination?
    Saurabh.

    Hi Saurabh,
    Table for Open Hub Services are :-
    RSBOHDEST : Directory of Open Hub Destinations
    RSBOHFIELDS : Fields of the Open Hub Destination
    Some FM are also available. It might be of help to you.
    RSB_API_OHS_DEST_GETLIST
    RSB_API_OHS_DEST_GETDETAIL
    Roy

  • Program or Function module to delete data from Open Hub Destination Table

    Hi All,
    Can anybody suggest me a Program or Function module to delete data from Open Hub Destination Table.
    Thanks & Regards,
    Vinay Kumar

    You can simply goto t-code SE14 mention the open hub destination table and Delete data by clicking on "Activate and Adjust database" with radio button "Delete Data".
    Regards,
    Arminder

  • Open hub destination as ECC table

    Hi All,
    Is it possible to create a open hub destination to an ECC table?
    Can only see the option for csv, tabke and thrird party.
    Thanks,
    Nick.

    Alright Arun,
    Yeah I was thinking of that.  If we have 100's of R3 tables where all the data needs extracting, then cleansing then imported to another R3 system would you consider doing this in BI or would you go R3  to R3???
    Cheers,
    Nick.

  • Open Hub Destination export to database table

    Hi Experts,
    I would like to create a multi source universe between SAP and non-SAP system. As we know the Jco connection to create a relational connection to access infocube data is not that great. I'm thinking or may be wrong in my approach that I would like to create a open hub destination of type database table on top of the infocube so that I'll have a new OHD table (/BIC/OH*) is created in BW database. Here comes my question that, can I query this new table in BW database via universe by creating a relational connection to the BW database server ?
    Please let me know your thoughts and suggestions!
    Thanks,
    Abhi 

    Hi
    I would like to create a multi-source universe between SAP and non-SAP system.
    Yes...In IDT...To build a universe on SAP NetWeaver BW, you must create a multisource-enabled data foundation based on a secured relational connection to SAP NetWeaver BW. Then build the business layer on this data foundation.
    I would like to create an open hub destination of type database table on top of the infocube so that I'll have a new OHD table (/BIC/OH*) is created in BW database.
    Creating universe on top of OHD tables is new idea, but while creating universe in IDT, we can able to select only below info providers.
    Info Object
    Info Cube
    DSO
    Multi Providers
    Virtual Info providers
    So we are not able to create universe on top of OHD tables. Please find the below screenshot.
    Here comes my question that, can I query this new table in BW database via universe by creating a relational connection to the BW database server ?
    No.....In 4.0 IDT,  OHD tables is not one of the info provider to select while creating universe. So it’s not possible.
    PLEASE mark the blog helpful & answered

  • Open Hub Destination Table Vs File

    Hello Experts,
    We have a requirement to send data from BW to 3rd party system every day. In the Open hub destination we have two options for the target.
    1. Table
    2. File
    ETL tool like Informatica will extract the data from table or file into 3rd party sytem.
    Which way is the best in performance and maintenance?
    Thanks in advance
    Sree

    Hi,
    If you follow the delta mechanism then table is good. You can also go for File not a problem. See the both eamples in the following articles.
    If you use Fiels then go for Application Server option, i.e. AL11 becasue no one can change teh files in this location, you can restrict the authorizations.
    Open Hub Destination: Part 1
    http://www.sdn.sap.com/irj/servlet/prt/portal/prtroot/com.sap.km.cm.docs/library/business-intelligence/m-o/open%20hub%20destination%3a%20part%201.pdf
    Open Hub Destination: Part 2
    http://www.sdn.sap.com/irj/servlet/prt/portal/prtroot/com.sap.km.cm.docs/library/business-intelligence/m-o/open%20hub%20destination%3a%20part%202.pdf
    Thanks
    Reddy

  • Help required for Creation of process chain for Open hub destination(table)

    Hi Experts,
    I am working on creating process chain for the open hub destination which is targeting data transfer into Table. Which is then push to SAP PI.
    2.) Steps to create a process chain for OHD????? Not knowing and even after that
    3.) What are the steps for Pushing it to PI (I mean we have generated client proxy also) But need to know ABAP related to trigger process chain
    please need your expertise and knowledge. Well atleast with the point 2 I need BI experts help...
    Regards,
    [Gaurav Patwari|http://gauravpatwari.wordpress.com]

    Hi Gaurav,
    Creating Process chain for OHD.
    1. Go to RSPC-> Create a astart variant.
    2. Click on process typesin the menu. Fourth from the left. Under this go to LOAD PROCESSES AND POST PROCESSING option.
    3. The fifth option is "DATA EXPORT INTO EXTERNAL SYSTEMS". Click on that.
    4. Click on F4 and choose your infospoke name.
    5. Press the RIGHT button.(if it asks for variant, create a new variant in the same small window besideswritting bar).
    6. Join the START process and the 2nd process via dragging.
    7. Now add your required other processes.
    The general process chain processes are :
    1. Start.
    2. Delete Indexes (if data loaded to cube).
    3. Load process.
    4. Delete duplicate request (if it is a full load in cube).
    5. Create Indexes.
    Hope it helps.
    Preet

  • Open Hub Destination table structure update

    Hi,
         I created a open Hub detsination and activation created the table  and populated the data in the table and have a view over the table.
         I need to add few more fields in the open hub and when I did that and activated again , it is supposed to change the table structure adding the new fields that were added in the open hub.
       The activation was successful but the table structure still stays the same.I didn't see the new fields in the generated table.
       Anyone having similar issues.
      Thanks in advance

    We deleted all the related objects from Staging and production and re transported them from Development as though they are being sent for the first time and that works.Tough this approach is not good we need to use it as I didn't find any other solution

  • Open Hub Destination question

    Hi,
    We want to be able to extract data from a cube and would like to use Open Hub Destination for it. We also want to extract deltas.
    The question is ..
    Should we be making a single info package for this with delta capability ..
    or 2 separate info packages .. 1 for full mode and another for delta mode.
    Regards
    Vandana

    Hi,
    Yes what you said it is write
    from which cube do u want to retract the daat that cube will be the source
    first create the open hub in RSBO tcode and
    in open hub destination select aht and create the transformation between the table( to where u want to export)
    then create the DTP between the cube to the table
    and directly u can use the delta DTP first all the records will come and later onle delta records will be updated.
    Thansk & Regards,
    sathish

  • Use of Open Hub Destination to load data into BPC

    Hi Gurus,
    I want to load the data from SAP BI to BPC (NW version), using Open Hub Destination. I want to know few things about this.
    1) What method should I use? Should I use Destination as Flat Files or Database tables? If Database tables, how can I use this data in the tables in BPC?
    2) If I go for Flat Files, which  is saved in the application server of BI, how can I import those files to BPC and use it using Data Manager?
    3) Also, in case of Flat Files, there are two files which are created. One is the control file and other one is the data file. How can I use both of them? Or can I just use data file without header?
    Your replies will be much appreciated.
    Thanks,
    Abhishek

    Hi Anjali,
    I can use the standard data manager package from BI to BPC, if the CSV file is available in the BPC Server or in case if I am directly extracting from BW object, InfoObject or InfoCube.
    But, since I will be using Open Hub, the output of this can be in a Database Table or a CSV file, preferably in SAP Application Server. In such cases, how can I use the Database table or CSV file in the standard Data Manager Package?
    Thanks for your reply.
    Abhishek

  • SAP BI 7.0 to SAP PI to FTP and with Open Hub Destination ...Help!!!!

    Dear SCN Experts,
    I am currently working on scenario where I have a requirement to push data from SAP BI 7.0 to SAP PI 7.
    And I am using Client proxy for SAP BI to send the data from BI tables to SAP PI and then to write to FTP address.
    Now the challenge I am facing is the ABAP Proxy how to use it with Process Chain and also I am having Open hub destination created for the same.(Specifically with the new version i.e. 7.0)
    Can you atleast make me understand what are the steps involved in this for Client Proxy, Process Chaing, How will proxy trigger and other regarding the same.
    I have searched SDN but got the document having older versions in it. Which doesn't serve the purpose.
    Regards,
    [Gaurav Patwari|http://gauravpatwari.wordpress.com]

    Hi Michal,
    Thanks for the reply buddy.
    I know that we can run the scheduled report for the proxy to fetch the report. But the client requirement is to use process chain, Open hub destination which fetches data from the ODS to Ztable created.
    We need to fetch that data from the table via our proxy. I am familiar with that report related method.
    I have one document using the same method but is using Infospoke(which is now obsolete) so have to OHD and also some of the proxy of XI's older version. So is not helping me out.
    Please do the needful. Or can you send me some sample scenario like this with screen shots. It will be a great help.
    Regards,
    [Gaurav Patwari|http://gauravpatwari.wordpress.com]

  • Open hub Destinations - Third-Party Tools As Destinations

    Dear all, below is the SAP Help process flow steps. I am quiet clear of all the steps except the APIs. Where do we define these APIs? When it says "You use API RSB_API_OHS_DEST_SETPARAMS to define the parameters for the third-party tool that are required for the extraction" where do I use it? Is it in the 3rd party tool? or is it in SAP BI.
    If it is in 3rd party tool, where in 3rd party tool. Do I need to write as code ? or would I have an option in the 3rd party tool?
    If someone can help me with Microsoft connector using Microsoft SQL Server 2008 Integration Services packages to SAP BI that would be helpful.
    Where in Microsoft SQL Server 2008 Integration Services packages can I use/put this APIs? Any help Plzz
    If it is in SAP BI where in SAP BI do I have to use these APIs? Any help is highly appreciated
    Regards,
    KK
    Process Flow:
    Extraction to the third-party tool can be executed as follows:
           1.      You define an open hub destination with Third-Party Tool as the destination type.
           2.      You create an RFC destination for your third-party tool and enter it in the definition of the open hub destination.
           3.      You use API RSB_API_OHS_DEST_SETPARAMS to define the parameters for the third-party tool that are required for the extraction.
           4.      You either start extraction immediately or include it in a process chain. You can also start this process chain from the third-party tool using process chain API RSPC_API_CHAIN_START. The extraction process then writes the data to a database table in the BI system.
           5.      When the extraction process is finished, the system sends a notification to the third-party tool using API RSB_API_OHS_3RDPARTY_NOTIFY.
           6.      The extracted data is read by API RSB_API_OHS_DEST_READ_DATA.
           7.      The status of the extraction is transferred to the monitor by API RSB_API_OHS_REQUEST_SETSTATUS.

    It's been a along time since i worked with open hub, but here goes:
    The Third Party tool needs parameters or switches to work correctly. Think of the old DOS commanda the  "/a" "/b" suffix.
    The APIs are simply a way to pass these parameters from the BW source to the Third Party target.
    If your tool does not need them, don't worry about them.
    John Hawk

  • Open Hub Destination-Third Party Tool

    dear all,
    can any one guide me as to how data(master data in particular) from bw can be transferred to oracle data base using open hub destiantion.is rfc connection between bw and third part tool enough,do we have to maintain the same structure in third party tool before executing DTP and how can this be done??
    i am now able to transfer data from bw into a flat file and also to data base (bw) and not aware of the this functionality-third party tool .any light thrown on this area is welcome and points will be awarded ,
    thanks ,
    sasidhar gunturu

    Some light at the end of tunnel for sure...
    Extraction to the third-party tool can be executed as follows:
           1.      You define an open hub destination with Third-Party Tool as the destination type.
           2.      You create an RFC destination for your third-party tool and enter it in the definition of the open hub destination.
           3.      You use API RSB_API_OHS_DEST_SETPARAMS to define the parameters for the third-party tool that are required for the extraction.
           4.      You either start extraction immediately or include it in a process chain. You can also start this process chain from the third-party tool using process chain API RSPC_API_CHAIN_START. The extraction process then writes the data to a database table in the BI system.
           5.      When the extraction process is finished, the system sends a notification to the third-party tool via API RSB_API_OHS_3RDPARTY_NOTIFY.
           6.      The extracted data is read by API RSB_API_OHS_DEST_READ_DATA.
           7.      The status of the extraction is transferred to the monitor by API RSB_API_OHS_REQUEST_SETSTATUS.
    For more details click below -
    http://help.sap.com/saphelp_nw2004s/helpdata/en/43/7a69d9f3897103e10000000a1553f7/content.htm
    Hope it Helps
    Chetan
    @CP..

  • Open Hub Destination problem

    Hi
    I have created open hub destination(OHD) on a cube and then getting file in CSV format on application server.
    I have to move this file to BPC but problem is wen run DTP in OHD it generated 2 files one with header and other with data
    How can i make it to generate one file with header and then data in it???
    How can i send this file automatically to BPC server( BPC is Microsoft version not netweaver one).

    Hi Sadik
    BPC is very new in market and there is not much help and yes it has base on MS SQL but thing is all data load is via CSV files only not thru direct table and stuff.
    I have tried making DB connect but i m not able to make RFC connection and there is no basis support I have in my project.
    I have solved my problem but there is a small stuck up. I am downloading data from internal table into file but this code is dumping file into my local machine. I want the file to be directly dumped into the application server.
    Following is the code:
    CALL FUNCTION 'MS_EXCEL_OLE_STANDARD_DAT'
    EXPORTING
      file_name = 'C:\Test' ????????? [This is the main problem]
    *file_name = 'DIR_HOME  E:\usr\sap\BDV\DVEBMGS30\work\ZBPC_1000' "set this for application server location
    CREATE_PIVOT = 0
    DATA_SHEET_NAME = ' '
    PIVOT_SHEET_NAME = ' '
    PASSWORD = ' '
    PASSWORD_OPTION = 0
    TABLES
    PIVOT_FIELD_TAB =
    data_tab = lt_data_f "internal table with data
    fieldnames = lt_header "internal table with header
    EXCEPTIONS
    file_not_exist = 1
    filename_expected = 2
    communication_error = 3
    ole_object_method_error = 4
    ole_object_property_error = 5
    invalid_filename = 6
    invalid_pivot_fields = 7
    download_problem = 8
    OTHERS = 9
    IF sy-subrc <> 0.
    MESSAGE ID sy-msgid TYPE sy-msgty NUMBER sy-msgno
    WITH sy-msgv1 sy-msgv2 sy-msgv3 sy-msgv4.
    ENDIF.
    Edited by: Navneet Thakur on Jul 2, 2009 2:53 PM

  • Populate text description in the open hub destination

    Hi SDNers,
    I have requirement with open hub destination, i am taking the cube data to SQL server database.
    in that scenario i need text description in the cube ,but we can not see the text description data in the cube(because cube can hold dimension data and utmost navigational attributes data but no textual description).
    Text data lies out side the cube but i need couples of text description for master data texts.
    Do i need to write the code to populate(look-up) the text description from text table.
    Can any one suggest an idea.
    Thanks,
    Satya

    Yes, you may want to write a code on transformations from Cube to Open hub to go and get text from Text table of that Info object.
    Make sure to create a new IO of lenght equal to Text field in IO and add this new IO in the Open hub and in transformations write a code at field  like:
    select single TXTSH from /BIC/Txxxxx into wa_text where
    key_of_p_table = /bic/source_field
    and langu = 'EN'
    If sy-subrc = 0.
    result = wa_text.
    endif.
    clear:wa_text

Maybe you are looking for

  • Office 2013 RMS Client Configuration - Your Machine is not setup for IRM

    Hi All, I have successfully setup a test AD RMS farm but stuck at trying to configure Office 2013 to work with this. The farm configuration is as below DC: Windows 2003  AD RMS Server: Windows 2008/ SQL Server 2008 R2 Client: Windows 7 64 bit with Of

  • No bootable device - Windows Vista

    Hi there. I've seen many posts with this question all over the internet, but I never got an answer, so I'm hoping someone here can help me. I turned off my computer last night and it was fine, but this morning when I turned it on the following messag

  • FAGLL03 : Submit syntax for dynamic selections

    Hi Experts, My z report contains following fields in selction screen. 1 . G/ L account 2. Comapny code 3. posting date 4. document type 5. layout In my z report i used following syntax for passing selection screen values to standard program and getti

  • Port in use

    I've got something wedged in a host or listener config. I had 11.2.0 running, scratched it off and to start over. I removed the software and reinstalled. Now when I run netca and try to create a listener on the port I was using it fails saying the po

  • Slow Solid State hard drive startup

    I just installed a 256gb ssd drive (from OWC) into my 2007 Santa Rosa 2.4 core 2 duo 17" macbook pro. I had done a clone from my old hard drive (7200rpm 320gb) while it was still in the mbp to the ssd drive in an external enclosure using Superduper.