Open hub destination of text and attribute for Functional Area

Hi Expert
I have a requirement to accommodate both attribute ( 0FUNCT_LOC_ATTR)  and text ( 0FUNCT_LOC_TEXT ) for functional area in open hub destination .For this a DSO has been created from both 0FUNCT_LOC_ATTR, 0FUNCT_LOC_TEXT data sources .Problem is that DSO has 0FUNCT_AREA ( Functional Area ) as only key fields . So text for both english ( EN ) and Dutch ( NL) for particular 0FUNCT_AREA not getting loaded .If we assign both 0LANGU and 0FUNCT_AREA as composite key of dso , data from attribute transformation is loaded as blank as there is no language filed in attribute data source ( 0FUNCT_LOC_ATTR ).   Please let me know how can I design to accommodate both text for different language and functional area for open hub destination  .
I know that I can create separate open hub destination for master data TEXT and Attribute.
Any help is highly appreciated.
Regards
Saikat

HI Saikat,
You can create an additional characteristics counter and make it a key field in the DSO.After that we can populate this characteristics in the end routine .It value would increase by 1 everytime  different text comes for the same functional area.
Sample code
data: counter type n,
         w_farea type /bio/oifunc_area.
Sort result_package by func_area.
count =0.
*checking whether we have the same functional area or we have a different functional area.
Loop at result_package assigning <Result_fields>.
if w_farea is initial or w_farea = <result_fields>-func_area.
count = count +1.
else
count =1.
endif.
w_farea= <result_fields>-func_area.
<Result_fields>-zcount = count.
ENDLOOP.
Clear w_farea.
This way the value of counter would get increased by one everytime a new text comes for the same functional area.We can have any number of texts for the same functional area by this method.
The only issue would be when some delta update comes for the existing text in this method.If you are doing a full load to the DSO from the datasource, then there would be no issue with this method.

Similar Messages

  • Issue for DTP from DSO to open hub destination

    Hello Gurus,
            I have a issue for DTP from DSO to open hub destination, long text for error in the monitor is as follows:
              " Could not open file
    SAPBITFS\tst_bi\bit\BIWork\GENLGR_OneTimeVendr_2 on application server"
              " Error while updating to target ZFIGLH03 (type Open Hub Destination)     "
          for open hub destination, I check the configure for logical file name , which is "
    SAPBITFS\tst_bi\bit\BIWork\GENLGR_OneTimeVendr",
    I am wondering where that file " 
    SAPBITFS\tst_bi\bit\BIWork\GENLGR_OneTimeVendr_2" in the error message comes from?
    Many thanks,

    Hi
    You do not need to create a file in application server. It will be created automatically.
    But if you have defined a logical file name in tcode FILE and used that in OHD and if it is not correct then it will show a conflict. Check this out.

  • Text Values in Open Hub Destination

    Hi ,
      Can anybody tell how to extract the text of particular info object using open hub destination into a file.
      For ex data is loaded from sales billing cube, ship to customer code is extracted , but need ship to customer name . how to extract this in open hub destination using BI 7.
    In queries , we use display key and text then get both key and text values. similar like that is it possible to get in open hub or not.
    Thanks .

    Hi Uppala,
    Following are 2 approaches to achieve this
    Approach 1:
    1. Create info spoke which will extract data from cube to data base table (do not use file)
    2. Create one more info spoke which will extract master data text to data base table (again do not use file)
    3. Create join on both the data base tables.
    4. Write a ABAP program which will create a file based on the joins of tables.
    Approach 2 :
    1. Create multiprovider on cube and master data info object
    2. Create info spoke based on the multi provider, that will extract data to flat file
    For 2nd approach u have to put certain filters to avoid unwanted rows
    Regards,
    Ajinkya

  • Open Hub Destination question

    Hi,
    We want to be able to extract data from a cube and would like to use Open Hub Destination for it. We also want to extract deltas.
    The question is ..
    Should we be making a single info package for this with delta capability ..
    or 2 separate info packages .. 1 for full mode and another for delta mode.
    Regards
    Vandana

    Hi,
    Yes what you said it is write
    from which cube do u want to retract the daat that cube will be the source
    first create the open hub in RSBO tcode and
    in open hub destination select aht and create the transformation between the table( to where u want to export)
    then create the DTP between the cube to the table
    and directly u can use the delta DTP first all the records will come and later onle delta records will be updated.
    Thansk & Regards,
    sathish

  • Open Hub Destination Table Vs File

    Hello Experts,
    We have a requirement to send data from BW to 3rd party system every day. In the Open hub destination we have two options for the target.
    1. Table
    2. File
    ETL tool like Informatica will extract the data from table or file into 3rd party sytem.
    Which way is the best in performance and maintenance?
    Thanks in advance
    Sree

    Hi,
    If you follow the delta mechanism then table is good. You can also go for File not a problem. See the both eamples in the following articles.
    If you use Fiels then go for Application Server option, i.e. AL11 becasue no one can change teh files in this location, you can restrict the authorizations.
    Open Hub Destination: Part 1
    http://www.sdn.sap.com/irj/servlet/prt/portal/prtroot/com.sap.km.cm.docs/library/business-intelligence/m-o/open%20hub%20destination%3a%20part%201.pdf
    Open Hub Destination: Part 2
    http://www.sdn.sap.com/irj/servlet/prt/portal/prtroot/com.sap.km.cm.docs/library/business-intelligence/m-o/open%20hub%20destination%3a%20part%202.pdf
    Thanks
    Reddy

  • Guide on Open hub destinations

    Hello Experts,
    We will be extracting and transporting the data from BW and other systems into the Oracle Database. Can anyone plz let me know the idea on the open hub destinations in detail...

    for 7.0:
    http://help.sap.com/saphelp_nw70/helpdata/en/b2/e50138fede083de10000009b38f8cf/frameset.htm
    for 3.x:
    http://help.sap.com/saphelp_nw04/helpdata/en/b2/e50138fede083de10000009b38f8cf/frameset.htm
    please search the forum.

  • Populate text description in the open hub destination

    Hi SDNers,
    I have requirement with open hub destination, i am taking the cube data to SQL server database.
    in that scenario i need text description in the cube ,but we can not see the text description data in the cube(because cube can hold dimension data and utmost navigational attributes data but no textual description).
    Text data lies out side the cube but i need couples of text description for master data texts.
    Do i need to write the code to populate(look-up) the text description from text table.
    Can any one suggest an idea.
    Thanks,
    Satya

    Yes, you may want to write a code on transformations from Cube to Open hub to go and get text from Text table of that Info object.
    Make sure to create a new IO of lenght equal to Text field in IO and add this new IO in the Open hub and in transformations write a code at field  like:
    select single TXTSH from /BIC/Txxxxx into wa_text where
    key_of_p_table = /bic/source_field
    and langu = 'EN'
    If sy-subrc = 0.
    result = wa_text.
    endif.
    clear:wa_text

  • DTP Delta request for open hub destination.

    Hi expert,
    As you know, In BW 7.0, open hub destination is used for open hub service, and you can create tranformation/DTP for the destination. and in some case the DTP delta is enabled.
    my questions are
    1) in the context menu of open hub destination, I did not find something like manage, which could list all the DTP request executed for this open hub destination..but for normal infocube or ods, you can find a list.
    2) if the DTP could enable the delta, where can I delete the latest delta request, and have a new load..
    3) I assume the DTP could be excuted with batch integrated with process chain by process type 'DTP', right ? if the destination type is third party tool, can it also be integrated with process chain ?
    4) can give some obvious difference of InfoSpoke and open hub destination in terms of advantage and disadvantage..
    Thanks a lot for the feedback.
    Best Regards,
    Bin

    Hi ,
    Please look at the links below . hope it helps .
    http://www.sdn.sap.com/irj/scn/go/portal/prtroot/docs/library/uuid/d063282b-937d-2b10-d9ae-9f93a7932403?QuickLink=index&overridelayout=true
    http://www.sdn.sap.com/irj/scn/go/portal/prtroot/docs/library/uuid/e830a690-0201-0010-ac86-9689620a8bc9?QuickLink=index&overridelayout=true
    Regards
    Santosh

  • SAP BI 7.0 to SAP PI to FTP and with Open Hub Destination ...Help!!!!

    Dear SCN Experts,
    I am currently working on scenario where I have a requirement to push data from SAP BI 7.0 to SAP PI 7.
    And I am using Client proxy for SAP BI to send the data from BI tables to SAP PI and then to write to FTP address.
    Now the challenge I am facing is the ABAP Proxy how to use it with Process Chain and also I am having Open hub destination created for the same.(Specifically with the new version i.e. 7.0)
    Can you atleast make me understand what are the steps involved in this for Client Proxy, Process Chaing, How will proxy trigger and other regarding the same.
    I have searched SDN but got the document having older versions in it. Which doesn't serve the purpose.
    Regards,
    [Gaurav Patwari|http://gauravpatwari.wordpress.com]

    Hi Michal,
    Thanks for the reply buddy.
    I know that we can run the scheduled report for the proxy to fetch the report. But the client requirement is to use process chain, Open hub destination which fetches data from the ODS to Ztable created.
    We need to fetch that data from the table via our proxy. I am familiar with that report related method.
    I have one document using the same method but is using Infospoke(which is now obsolete) so have to OHD and also some of the proxy of XI's older version. So is not helping me out.
    Please do the needful. Or can you send me some sample scenario like this with screen shots. It will be a great help.
    Regards,
    [Gaurav Patwari|http://gauravpatwari.wordpress.com]

  • Inconsistency with ZPMSENDSTATUS for Open Hub Destination

    Hi Experts,
    I dont know you have come accross this scenario but I will apprecaite all your advice.
    We have APO 7.0 and we are using the BW component to extract data to informatica.
    Our environment is as follow
    SAP SCM 7.0 SP8
    Informatica PWCD 8.1.6 hot fix 13
    1.we have created 5 chains with similar variants for ZPMSENDSTATUS
    2. Each variant relates to a folder and workflow to be executed within informatica
    3. When we execute the chains only two work correctly all the time and 3 fail
    the 3 that are failing are failing within the ZPMSENDSTATUS when running API RSB_API_OHS_DEST_SETPARAMS. This is odd since the other two only differ on the workflow to be executed and obviously on the Open Hub destination. Can anyone provide some advice on this?
    thanks,
    Nat

    Hi Maruthi,
    thanks for your response but here are the details for the variants that are working and failining
    working:
    DEST: < here we have the RFC that relates to the logical system and the program id which is also included in saprfc.ini>
    INFPARA: AERO/APO:WF_SAP_APO_PART_PLATFORM_ABP2_BGA:S_M_SAP_APO_PART_PLATFORM_ABP2
    CONTEXT: OHS API
    OHDEST: <OHDS defined as target>
    Failing:
    DEST: < here we have the RFC that relates to the logical system and the program id which is also included in saprfc.ini>
    INFPARA: AERO/APO:WF_SAP_APO_PART_PLATFORM_ABP4_ATR:S_M_SAP_APO_PART_PLATFORM_ABP4
    CONTEXT: OHS API
    OHDEST: <OHDS defined as target>
    So you see, they all target the same folder only different workflows. Still it works for the first and not for the second.

  • What is source system and open hub destination?

    Hello,
             Kindly tell me In BI data modeling workbench ,in modeling there is are two areas name as "Source system" and "Open Hub destination".What is mean by this???How we use it in BI?

    Hi,
    Source system as the name suggests mentions the source of data. The place from where you are extractin data to BW.
    http://help.sap.com/saphelp_nw2004s/helpdata/en/7a/27bcf087c7464db8b95eaa717b6e6a/content.htm
    OpenHub is for exporting data out of BW.
    https://www.sdn.sap.com/irj/sdn/go/portal/prtroot/docs/library/uuid/11e1b990-0201-0010-bf9a-bf9d0ca791b0
    https://www.sdn.sap.com/irj/sdn/go/portal/prtroot/docs/library/uuid/3042c5fc-21a8-2910-c79e-ad530260ae2e
    https://www.sdn.sap.com/irj/sdn/go/portal/prtroot/docs/library/uuid/01d3a090-0201-0010-9783-bc
    http://help.sap.com/saphelp_erp2005vp/helpdata/en/bf/50453c01f4f75fe10000000a11402f/frameset.htm
    Thanks,
    JituK

  • Multiprovider as Datasource for Open Hub Destination

    Hello,
    Can you pls let me know if we can used the Multiprovider as datasource for the Open Hub destination. Currently we are Bi7.0 and SP9. I am not able to see the option of Mutiprovider in the template.
    Appreciate the input ASAP.
    Thanks
    Gopal

    Multiprovider is not yet supported for Open Hub destination and hence you dont see that option.
    As of SAP NetWeaver 2004s SPS 6, the open hub destination has its own maintenance interface and can be connected to the data transfer process as an independent object. As a result, all data transfer process services for the open hub destination can be used.  You can now select an open hub destination as a target in a data transfer process.  In this way, the data is transformed as are all other BI objects.
    In addition to the InfoCube, InfoObject, and DataStore object, you can also use DataSources and InfoSources as templates for the field definitions of the open hub destination.
    The open hub destination now has its own tree under Modeling in the Data Warehousing Workbench. This tree is structured by InfoAreas.
    <b>Restrictions -</b>
    It is not currently possible to use MultiProviders as the source for the open hub destination.
    http://help.sap.com/saphelp_nw2004s/helpdata/en/43/7c836c907c7103e10000000a1553f7/content.htm
    Hope it Helps
    Chetan
    @CP..

  • What's the difference between Infospokes and Open Hub Destination in BI 7?

    I understand the API that a 3rd party tool uses to get data our of BW using infospokes.    Is it the same API that a 3rd party tools such as MS-SSIS uses?   Supposedly  that software only communicates with an OH Destination and not Infospokes.   So apparently is does use a different API.   If that's the case, where can I find documentation on that API for OH Destintations?

    Open Hub Destination (OHD) replaced InfoSpokes as of BI 7 SPS 6, for Open Hub Service. Existing InfoSpokes can be used still, but it's recommended that any new retractions be created as an Open Hub Desitnation, because it exposes more functionality (e.g. Transformations and functionalities with DTPs).
    Here's a document on [Open Hub Service - 3rd Party Integration|http://www.sdn.sap.com/irj/scn/go/portal/prtroot/com.sap.km.cm.docs/library/icc/bw-ohs%20open%20hub%20service%2c%20third-party%20integration.pdf] (see Page 7). Hope this helps.
    Edited by: Dennis Scoville on Aug 3, 2009 1:45 PM

  • Difference bw open hub destination and info spoke ???

    Hi,
    what are difference b/w open hub destination and info spoke ???
    Thanks In advance.
    Regards
      Nitin

    Hi,
    The open hub destination is the object that allows you to distribute data from a BI system to non-SAP data marts, analytical applications, and other applications. It ensures controlled distribution across multiple systems.
    The open hub destination defines the target to which the data is transferred.
    An InfoSpoke can extract data in two ways:
    1.Full mode (F): the data corresponding to the selection criteria of the InfoSpoke is transferred from the data source into the respective destination.
    2.Delta mode (D): records that were added since the last extraction only are transferred in this mode. This is only possible for the source objects InfoCube and DataStore object.
    See the below Articles:
    InfoSpoke: Part -1
    www.sdn.sap.com/irj/servlet/prt/portal/prtroot/com.sap.km.cm.docs/library/business-intelligence/g-i/infospoke%20part%201.pdf
    Open Hub Destination: Part 1
    http://www.sdn.sap.com/irj/servlet/prt/portal/prtroot/com.sap.km.cm.docs/library/business-intelligence/m-o/open%20hub%20destination%3a%20part%201.pdf
    Open Hub Destination: Part 2
    http://www.sdn.sap.com/irj/servlet/prt/portal/prtroot/com.sap.km.cm.docs/library/business-intelligence/m-o/open%20hub%20destination%3a%20part%202.pdf
    Thanks
    Reddy

  • Help required for Creation of process chain for Open hub destination(table)

    Hi Experts,
    I am working on creating process chain for the open hub destination which is targeting data transfer into Table. Which is then push to SAP PI.
    2.) Steps to create a process chain for OHD????? Not knowing and even after that
    3.) What are the steps for Pushing it to PI (I mean we have generated client proxy also) But need to know ABAP related to trigger process chain
    please need your expertise and knowledge. Well atleast with the point 2 I need BI experts help...
    Regards,
    [Gaurav Patwari|http://gauravpatwari.wordpress.com]

    Hi Gaurav,
    Creating Process chain for OHD.
    1. Go to RSPC-> Create a astart variant.
    2. Click on process typesin the menu. Fourth from the left. Under this go to LOAD PROCESSES AND POST PROCESSING option.
    3. The fifth option is "DATA EXPORT INTO EXTERNAL SYSTEMS". Click on that.
    4. Click on F4 and choose your infospoke name.
    5. Press the RIGHT button.(if it asks for variant, create a new variant in the same small window besideswritting bar).
    6. Join the START process and the 2nd process via dragging.
    7. Now add your required other processes.
    The general process chain processes are :
    1. Start.
    2. Delete Indexes (if data loaded to cube).
    3. Load process.
    4. Delete duplicate request (if it is a full load in cube).
    5. Create Indexes.
    Hope it helps.
    Preet

Maybe you are looking for

  • IMac G5 with Built-In iSight - System Freezes

    Hi, I have the new model iMac G5 that is now consistantly freezing. When I say freezing, I get the spinning wheel and nothing is accessable. I have to hold down the power button to reset it. I have called Apple and reset NVRAM, clear the cache's, etc

  • Errors with Blob

    Hi, I'm working with servlets and jsp's but my error occurs at the Blob object level. In my jsp I have an <img> tag and the src="" of that image(which will be binary data) will be retrieved from a servlet that recieves a Blob(the image) from the sess

  • Easy, ReadSoft and Seal Systems

    Hi all, Can any one explain or provide links to what these THIRD PARTY TOOLS pertain to and how integration can be done with SAP. What are these tools that are SAP certified which are used for Document Management System & Records and Case Management.

  • Automatic Login Trouble and Short Name Trouble

    Hello everyone, I just got a fresh install of Leopard going on my Macbook (wiped the hard drive during installation instead of just upgrading the OS), but I'm running into a brick wall (two of them, actually). First off, for some reason, my Macbook r

  • Thanks to Mandeep5062 Acrobat Pro download issue with Safari

    Before your response arrived I shut down Safari, loaded Firefox, went directly to the Adobe download page and attempted a download. The download started correctly and then switched over to the Creative Cloud downloader and I thought, "Here we go agai