Data Extraction in Open Hub Destination using Process Chain

Hi
I want to extract data in Open Hub Destination (database table)from Data Store Object(ODS) through Process Chain.
When i tried to create process chain, i found only one option under Process types - Data export into External System. Here it asks for Infospoke instead of Open Hub Destination.

Michael is correct and below is the rationale...
<a href="http://help.sap.com/saphelp_nw2004s/helpdata/en/43/58e1cdbed430d9e10000000a11466f/content.htm">http://help.sap.com/saphelp_nw2004s/helpdata/en/43/58e1cdbed430d9e10000000a11466f/content.htm</a>
<b>Integration</b>
You can use the data transfer process to update data to the open hub destination. The data is transformed in this process. Not all rules types are available in the transformation for an open hub destination: Formulas, reading master data, time conversion, currency translation and unit conversion are not available.
Hope it Helps
Chetan
@CP..

Similar Messages

  • Multiple data sources to open hub destination ??

    Hi,
            Is it possible to assign multiple data sources to open hub destination?
    At present we are FTP'ing a file to users. Users need one more field added to this file. The information for this field belongs to another ODS. I  checked the infospoke but it does not accept multiple data sources.
    Any ideas?
    Thanks!
    Edited by: BI Quest on Sep 4, 2008 8:12 PM

    You can create multiple transformations for open hub destination. Create destination of open hub destination to be file on application server and then use transfer mechanism to transfer files to desired destination.
    Regards,
    Pravin Karkhanis.

  • Open Hub 3rd Party - Process Chain Automation

    Hi,
    I have a requirement to fulfill the Open Hub extraction to 3rd party SQL Server DB using process chain automation.
    There are at times no deltas and during the execution of DTP since there are no delta's the request status shows GREEN but overall status as YELLOW because there are no new records.
    Until we clear this request to GREEN, we cant trigger the next load. I am doing these steps manually by going to tables/ FM's to change the status.
    I am looking for options for automation in process chain can someone help me on this request. Appreciate your help!
    Thanks,
    Pandu.

    Do you know when the delta will being zero records? Also when you say the overall status is yellow which status are you talking about?
    You can always use a ABAP program in your PC which will check for the latest request ID has records and turn the the status green so that dependent  data triggers are sent
    Thanks
    Abhishek Shanbhogue

  • Master Data Extraction through Open Hub services(InfoSpoke)

    Hello,
             I need your help on this issue,
    Here is my scenario...
    I want to extract Master data from BW (Info Object) as a file and put it in a server then later that can be picked up by some other deprtment for their process. This process has to be automated to run on a weekly basis.
    I know this can be done through a InfoSpoke(either ATTR or Text) and Process Chain but I want both attributes & text data in a single file because through InfoSpoke we can get either ATTR or TEXT data in a file.
    1)Is there any way we can have both TEXT and ATTRIBUTEs data in a single file.
    2) Also what are the options to transfer the file from a BW app server into another non-BW server?
    Currenty I created a BEx query against a InfoObject, ruuning a report and sending them the file manually. I want to automate the process.
    I would appreciate if you could give me input on the above.
    Thanks
    BI Consultant

    Hi,
    The simpleset way to do is to create a ODS with the required columns and map it to your TEXT and ATTR. Then you could use either a Query or a Info Spoke to extract data how ever you want it. Since you are using master data it would have been validated, so you don't need to use a reportable ODS, which takes time to activate the data. So create a info set on top of the ODS and do a simple query if you want to use it. If you want to automate completely and deliver the output as .CSV file to the destination server you may have to do little bit more and that will involve some serious ABAP work. The simpleset would be have a ODS 'Non reportable' and extract the data using info spoke.
    Let me know if you need more info.
    Thanks,
    Alex.

  • Save a csv file on local host, executing a DTP Open hub in a Process chain?

    Hi All,
    my client has asked me to try if there is a way to save in the local host c:\  a .CSV file generated from an Open HUB executed by a DTP in a Process chain.
    when i execute my DTP it work correctly and it save the .csv file on my c:\ directory.
    my client doesn't want to give th users authtorization on RSA1 and want to try if there is a way to put the DTP in a process chain
    and when executing the process chain every user can have the file on his own local c:\ directory.
    i tried this solution but it doesn't work i get an error on executin the process chain:
    Runtime Errors         OBJECTS_OBJREF_NOT_ASSIGNED
    Except.                     CX_SY_REF_IS_INITIAL
    Error analysis
        An exception occurred that is explained in detail below.
        The exception, which is assigned to class 'CX_SY_REF_IS_INITIAL', was not
         caught in
        procedure "FILE_DELETE" "(METHOD)", nor was it propagated by a RAISING clause.
        Since the caller of the procedure could not have anticipated that the
        exception would occur, the current program is terminated.
        The reason for the exception is:
        You attempted to use a 'NULL' object reference (points to 'nothing')
        access a component (variable: "CL_GUI_FRONTEND_SERVICES=>HANDLE").
        An object reference must point to an object (an instance of a class)
        before it can be used to access components.
        Either the reference was never set or it was set to 'NULL' using the
        CLEAR statement.
    can you give some advices please?
    Thanks for All
    Bilal

    Hi Bilal,
    Unfortunately, DTPs belonging to Open Hubs wich are targeted to a local workstation file, can't be executed through a process chain. 
    The only way of including such DTP in a process chain is changing the Open Hub so that it writes the output file in the application server.  Then, you can retrieve the file -through FTP or any other means- from the application server to the local workstation.
    Hope this helps.
    King regards,
    Maximiliano

  • Erase data from open hub destination data service

    Hi,
    I'm getting an error trying to delete data from a cube, for which data are extracted using open hub service and loaded into a table.
    The error given is the following:
    Request updated in 1794 already by 1795 target DTPR_4EEX9CGGXLTU2QIPWUAIEUDZK DTP request (ZRAPPEL)
    Message no. RSM077
    Diagnosis
    One or more DTP requests have already updated the request.
    Procedure
    Firstly, delete all update requests from your targets. Then you can delete the source request.
    Zrappel data is the open hub destination. But it is clear where requests for this destination, to let me delete the cube data.
    Thanks.

    Hi,
    Procedure
    Firstly, delete all update requests from your targets. Then you can delete the source request.
    Follow the same steps above , that means delete the requests from the Cube Zrappel.
    right click on the cube -
    > Delete Data
    now the syatem will promt you asking
    Do you want to delete the content of data target Zrapple ?
    say yes for it .
    hope this will solve your problem
    santosh

  • To view the data in open hub destination

    Hi all,
    Can anyone help me out how to check the data in open hub destination file.
    These are the details:
    Destination: File
    Type of file : Logical file
    Application Server file name : XXXXXXX
    If it was file name, i know it has to be checked in AL11 with directory name and file name.
    Here its a logical file name. Which is the directory corresponding to this?
    Thanks,
    tinkugeo

    Hi,
    Question: "Are all the logical files are saved in the home directory DIR_HOME?
                      Please correct me if I am wrong"
    Ans: When you create an Open hub destination using Physical file, the file will get created in DIR_HOME directory (by default) in AL11
    And for locating the physical path in AL11. Go through the below link. it gives ans to all your ques
    http://bx.businessweek.com/sap/view?url=http%3A%2F%2Fweblogs.sdn.sap.com%2Fpub%2Fwlg%2F17370
    Edited by: kumkum basu on May 20, 2010 8:20 AM

  • Open hub destination issue

    Hi,
    In our project we have Client Instance (eg: BWPD) and Application server (eg: BWPRD) defined for load balancing.
    We have created open hub and assigned destination server BWPD.
    When i execute the DTP manually in BWPD, it runs succesfully.
    However, the same DTP when placed in the process chain fails with error message :
    No Such File or Directory
    Could not open file D:\usr\sap\BWPD\A01\work\Material on application server
    Error while updating to target ZXXXX.
    Options Tried:
    Schedule process chain in the background server BWPD (same server which has been mentioned in Open hub dest) still DTP failed.
    Tried with Application server it failed.
    Tried with HOST as option it failed.
    couldn't make out what is going wrong. Any thoughts ?
    Regards.

    Hi there,
    found that doc quite useful.
    Maybe could shed some light to your issue.
    [Creating  Open Hub Destination  using a Logical file to extract the data|Creating  Open Hub Destination  using a Logical file to extract the data]
    Also, what OS do you have?
    is the Syntax Group accordingly created ?

  • Open Hub Destination with Application Server Issue

    Hi all,
    I have requriment that i need to save my monthly data in application server with the help of OPen Hub Destination.I have created an DSO
    on that DSO i have created an Open HUb with application server as target.When i store my every month file in application server that file name should be a per the system date.
    So i have Created an LOgical File----ztest_data
    Assgined Physical path to Logical File-----/usr/sap/<sysid>........
    created an Logical File name------Ztest_data_file....
    IN the open HUb i have selected Aplication Server as My target and i have checked the box.
    I have selected Logical FIle name and given the file name which i have created-----ztest_data_file...
    Runned DTP it is showing no errors,but when i check in AL11 ,in the path which i have given(/usr/sap/<sysyid>........,i couldnt find the file.....
    Please check this and let me know...
    Thanks In Advance,
    Bobby.

    Hi,
    Please check the link /people/jyothi.velpula/blog/2010/01/20/creating-open-hub-destination-using-a-logical-file-to-extract-the-data .
    Also is your DTP full or delta. Try running the DTP as Full.
    Hope it helps.
    Best Regards,
    Kush Kashyap

  • Text Values in Open Hub Destination

    Hi ,
      Can anybody tell how to extract the text of particular info object using open hub destination into a file.
      For ex data is loaded from sales billing cube, ship to customer code is extracted , but need ship to customer name . how to extract this in open hub destination using BI 7.
    In queries , we use display key and text then get both key and text values. similar like that is it possible to get in open hub or not.
    Thanks .

    Hi Uppala,
    Following are 2 approaches to achieve this
    Approach 1:
    1. Create info spoke which will extract data from cube to data base table (do not use file)
    2. Create one more info spoke which will extract master data text to data base table (again do not use file)
    3. Create join on both the data base tables.
    4. Write a ABAP program which will create a file based on the joins of tables.
    Approach 2 :
    1. Create multiprovider on cube and master data info object
    2. Create info spoke based on the multi provider, that will extract data to flat file
    For 2nd approach u have to put certain filters to avoid unwanted rows
    Regards,
    Ajinkya

  • Initial full load of Master data using process chain

    Hi All,
    Could you please help me regarding, initial master data load to characteristics with attributes and text. I need to load master data to 23 info objects, by using process chain can I do full load of master data to all info objects at a time. And one more doubt is, as per my knowledge we can't maintain more than one variant in an info package, is that right ? or we can ?
    Means Start Variant -> Info Package (0Customer_Text, 0Customer_Attr,0BILL_TYPE_TEXT, BILL_CAT_TEXT) -> DTP ( ", ", ", ") -> ACR.
    Your Help will be appreciated.
    Thanks & Regards
    Sunil

    Hi,
    "I need to load master data to 23 info objects, by using process chain can I do full load of master data to all info objects at a time."
    if there is no dependency between attributes then you add you can create process chains and trigger them at a time. No issues.
    we can't maintain more than one variant in an info package, is that right ? or we can ?
    With one info pack you can't load data to all 23 psa. because each data source have own psa. you need to sue 23 info packs.
    in general start variant--> info pack --> dtp (assuming as your bw 7.x)---> attribute change run.
    like that you need to create 23 chains 
    Or create one two big chains.
    one is for attribute and another for text.
    In attribute
    start varaint--> info pack(info bject 1)--DTP(infoobject 1))--> info pack(infoo bject 2)-->dtp(infoobject 2).
    Like that way you can create in series and parallel chains to load attributes data into info objects. at end you add change run for 6 info objects each. SAme you can do for text loads also.
    Thanks

  • Loading from flat file to dso using process chains

    hi,
    i am using BI7.0
    i am new to  process chains
    can anyone  explain how to load data from flat file to dso using process chains(i have created all the objects created) preffered if explained with an example

    You can find a lot info if you can searh SDN.
    Metachain
    Steps for Metachain :
    1. Start ( In this variant set ur schedule times for this metachain )
    2.Local Process Chain 1 ( Say its a master data process chain - Get into the start variant of this chain ( Sub chain - like any other chain ) and check the second radio button " Start using metachain or API " )
    3.Local Process Chain 2 ( Say its a transaction data process chain do the same as in step 2 )
    Steps for Process Chains in BI 7.0 for a Cube.
    1. Start
    2. Execute Infopackage
    3. Delete Indexes for Cube
    4.Execute DTP
    5. Create Indexes for Cube
    For DSO
    1. Start
    2. Execute Infopackage
    3. Execute DTP
    5. Activate DSO
    For an IO
    1. Start
    2.Execute infopackage
    3.Execute DTP
    4.Attribute Change Run
    Data to Cube thru a DSO
    1. Start
    2. Execute Infopackage ( loads till psa )
    3.Execute DTP ( to load DSO frm PSA )
    4.Activate DSO
    5.Delete Indexes for Cube
    6.Execute DTP ( to load Cube frm DSO )
    7.Create Indexes for Cube
    3.X
    Master loading ( Attr, Text, Hierarchies )
    Steps :
    1.Start
    2. Execute Infopackage ( say if you are loading 2 IO's just have them all parallel )
    3.You might want to load in seq - Attributes - Texts - Hierarchies
    4.And ( Connecting all Infopackages )
    5.Attribute Change Run ( add all relevant IO's ).
    Start
    Infopackge1A(Attr)|Infopackge2A(Attr)
    Infopackge1B(Txts)|Infopackge2B(Txts)
    /_____________________|
    Infopackge1C(Txts)______|
    \_____________________|
    ___________________|
    __\___________________|
    ___\__________________|
    ______ And Processer_ ( Connect Infopackge1C & Infopackge2B )
    __________|__________
    Attribute Change Run ( Add Infobject 1 & Infoobject 2 to this variant )
    1. Start
    2. Delete Indexes for Cube
    3. Execute Infopackage
    4.Create Indexes for Cube
    For DSO
    1. Start
    2. Execute Infopackage
    3. Activate DSO
    For an IO
    1.Start
    2.Execute infopackage
    3.Attribute Change Run
    Data to Cube thru a DSO
    1. Start
    2. Execute Infopackage
    3.Activate DSO
    4.Delete Indexes for Cube
    5.Execute Infopackage
    6.Create Indexes for Cube
    Some Links
    http://help.sap.com/saphelp_nw2004s/helpdata/en/8f/c08b3baaa59649e10000000a11402f/frameset.htm
    https://www.sdn.sap.com/irj/servlet/prt/portal/prtroot/docs/library/uuid/8da0cd90-0201-0010-2d9a-abab69f10045
    https://www.sdn.sap.com/irj/servlet/prt/portal/prtroot/docs/library/uuid/19683495-0501-0010-4381-b31db6ece1e9
    https://www.sdn.sap.com/irj/servlet/prt/portal/prtroot/docs/library/uuid/36693695-0501-0010-698a-a015c6aac9e1
    https://www.sdn.sap.com/irj/servlet/prt/portal/prtroot/docs/library/uuid/9936e790-0201-0010-f185-89d0377639db
    https://www.sdn.sap.com/irj/servlet/prt/portal/prtroot/docs/library/uuid/3507aa90-0201-0010-6891-d7df8c4722f7
    https://www.sdn.sap.com/irj/servlet/prt/portal/prtroot/docs/library/uuid/263de690-0201-0010-bc9f-b65b3e7ba11c

  • Hierarchy Extraction using Open hub Destination

    Hi,
    I need to extract hierarchy data for Application code(YAPPLCD) using OHD. In the Hierarchy there are 4 components(Industry, Market, Application Group and Application Code). i need to get all the 3 details for the Application Code(i.e to which Indusrty does it belong, Market and Application Group).But HIerarchy table is having one record for each apllication code, Market, Application Group, Application code.
    How to get relevant data for the application code?
    My file format should be:
    Appl.Code | Appl.Group | Market | Industry.

    Hi Suma...
    The SAP Open Hub service enables you to distribute data from an SAP BI/BW system into external data marts, analytical applications, and other applications. With this, you can ensure controlled distribution using several systems. The central object for the export of data is the InfoSpoke. Using this, you can define the object from which the data comes and into which target it is transferred, leveraging SQL Server Integration Services (SSIS) as the transfer/transform mechanism. 
    SAP BI objects such as InfoCubes, ODS objects, or InfoObjects can act as open hub data sources. You can select database tables or flat files as open hub destinations, and then use those destinations as actual sources within SSIS.  To automate the extraction using SAPu2019s Open Hub Service, youu2019ll need to setup a process chain. Note that both a full and delta modes are also available as the extraction mode.
    As stated already with direct extraction from SAP R/3, once data is extracted out of SAP NetWeaver® BI (SAP BW), the destination of that data could be any relational database (such as SQL Server), and then reports (from SSRS) and multi-dimensional cubes (from SSAS) can be created from that database data.  The resulting SQL Server reports/cubes could then be used within the Microsoft Business Intelligence delivery mechanisms, such as through Excel, SharePoint, PerformancePoint, etc.
    Hope this below link may helpful
    https://www.sdn.sap.com/irj/scn/go/portal/prtroot/docs/library/uuid/c08b1f70-6fb1-2b10-97a5-df4c503e03cb
    Regards
    Sudheer

  • Use of Open Hub Destination to load data into BPC

    Hi Gurus,
    I want to load the data from SAP BI to BPC (NW version), using Open Hub Destination. I want to know few things about this.
    1) What method should I use? Should I use Destination as Flat Files or Database tables? If Database tables, how can I use this data in the tables in BPC?
    2) If I go for Flat Files, which  is saved in the application server of BI, how can I import those files to BPC and use it using Data Manager?
    3) Also, in case of Flat Files, there are two files which are created. One is the control file and other one is the data file. How can I use both of them? Or can I just use data file without header?
    Your replies will be much appreciated.
    Thanks,
    Abhishek

    Hi Anjali,
    I can use the standard data manager package from BI to BPC, if the CSV file is available in the BPC Server or in case if I am directly extracting from BW object, InfoObject or InfoCube.
    But, since I will be using Open Hub, the output of this can be in a Database Table or a CSV file, preferably in SAP Application Server. In such cases, how can I use the Database table or CSV file in the standard Data Manager Package?
    Thanks for your reply.
    Abhishek

  • Data selection in new Open Hub destination Vs Infospoke

    In BW 3.5, we can use T code RSBO to create Infospoke, In the selection tab, we can use any infoobject in the source infocube as the data filterfor the infoSpoke, e.g I can set company code as 1002 etc, while in the new  BI 7.0, Open Hub Destination,  where I can achieve the same by set the filter for a specific infoobject in the source cude?
    Edited by: Jianbai on Mar 29, 2010 11:49 PM
    Edited by: Jianbai on Mar 29, 2010 11:51 PM

    For Open Hub in BI7 we are creating transformation and DTP. The same resrtiction you can give in DTP (activate it after change).

Maybe you are looking for

  • Using COUNTIF on time values in pop-up menu (employee schedule template)

    I have modified the Numbers employee schedule template to my liking and it works very well. I would like to add a column to the beginning which shows, at a glance, how many people I have scheduled for AM shifts and PM shifts on a given day. My AM shi

  • Synchronous AI, AO and DI

    Hello LabVIEW community - I am have some difficulty in getting my application to work properly - specifically monitoring the status of a DI line during AI and generating an AO waveform. The equipment I am using consists of: cDAQ 9172 chassis, 9401 DI

  • Lookup in a grid

    Hello all.. I have an entity object which represents a table of type: description(VARCHAR),date(Date), Id(Numeric). The Id is FK for table: Id(Numeric PK), Name VARCHAR Now in my application I have a gridControl which should represent table1 but inst

  • How can i use page break for every 5 records in sap scripts

    on every 5 lines of records i have go for a new page so what is procedure to do this .if possible send me with coding .

  • Automation Error - Can't register type library

    I am running Windows 7, 32-bit. I tried to add a reference from Visual Studio but Photoshop was not visible. So I decided to try to register the type library manually. I am trying to automate photoshop but I cannot get the type library registered. Th