RE: Open Hub Destination  "Indicator for Semantic Key"

Hi ..
While Creating OpenHub Destination In BI 7.O ,  In Fileds tab we can find "Indicator for Semantic Key**"  column Selection check box , can any one let me know Why we select this check box  ? what is it's purpose ?
Edited by: navenas on Feb 2, 2010 3:08 PM

HI,
Semantic key:
If you set the Semantic Key indicator, the system selects all the fields in the field list as semantic keys, if they are suitable. You can change this selection in the field list. However, note that duplicate records may result from using a semantic key.
Check this out for more details
http://help.sap.com/saphelp_nw70/helpdata/en/43/7b31c102a07100e10000000a1553f7/content.htm
regards,
mahesh

Similar Messages

  • SAP BI 7.0 to SAP PI to FTP and with Open Hub Destination ...Help!!!!

    Dear SCN Experts,
    I am currently working on scenario where I have a requirement to push data from SAP BI 7.0 to SAP PI 7.
    And I am using Client proxy for SAP BI to send the data from BI tables to SAP PI and then to write to FTP address.
    Now the challenge I am facing is the ABAP Proxy how to use it with Process Chain and also I am having Open hub destination created for the same.(Specifically with the new version i.e. 7.0)
    Can you atleast make me understand what are the steps involved in this for Client Proxy, Process Chaing, How will proxy trigger and other regarding the same.
    I have searched SDN but got the document having older versions in it. Which doesn't serve the purpose.
    Regards,
    [Gaurav Patwari|http://gauravpatwari.wordpress.com]

    Hi Michal,
    Thanks for the reply buddy.
    I know that we can run the scheduled report for the proxy to fetch the report. But the client requirement is to use process chain, Open hub destination which fetches data from the ODS to Ztable created.
    We need to fetch that data from the table via our proxy. I am familiar with that report related method.
    I have one document using the same method but is using Infospoke(which is now obsolete) so have to OHD and also some of the proxy of XI's older version. So is not helping me out.
    Please do the needful. Or can you send me some sample scenario like this with screen shots. It will be a great help.
    Regards,
    [Gaurav Patwari|http://gauravpatwari.wordpress.com]

  • Open hub destination of text and attribute for Functional Area

    Hi Expert
    I have a requirement to accommodate both attribute ( 0FUNCT_LOC_ATTR)  and text ( 0FUNCT_LOC_TEXT ) for functional area in open hub destination .For this a DSO has been created from both 0FUNCT_LOC_ATTR, 0FUNCT_LOC_TEXT data sources .Problem is that DSO has 0FUNCT_AREA ( Functional Area ) as only key fields . So text for both english ( EN ) and Dutch ( NL) for particular 0FUNCT_AREA not getting loaded .If we assign both 0LANGU and 0FUNCT_AREA as composite key of dso , data from attribute transformation is loaded as blank as there is no language filed in attribute data source ( 0FUNCT_LOC_ATTR ).   Please let me know how can I design to accommodate both text for different language and functional area for open hub destination  .
    I know that I can create separate open hub destination for master data TEXT and Attribute.
    Any help is highly appreciated.
    Regards
    Saikat

    HI Saikat,
    You can create an additional characteristics counter and make it a key field in the DSO.After that we can populate this characteristics in the end routine .It value would increase by 1 everytime  different text comes for the same functional area.
    Sample code
    data: counter type n,
             w_farea type /bio/oifunc_area.
    Sort result_package by func_area.
    count =0.
    *checking whether we have the same functional area or we have a different functional area.
    Loop at result_package assigning <Result_fields>.
    if w_farea is initial or w_farea = <result_fields>-func_area.
    count = count +1.
    else
    count =1.
    endif.
    w_farea= <result_fields>-func_area.
    <Result_fields>-zcount = count.
    ENDLOOP.
    Clear w_farea.
    This way the value of counter would get increased by one everytime a new text comes for the same functional area.We can have any number of texts for the same functional area by this method.
    The only issue would be when some delta update comes for the existing text in this method.If you are doing a full load to the DSO from the datasource, then there would be no issue with this method.

  • DTP Delta request for open hub destination.

    Hi expert,
    As you know, In BW 7.0, open hub destination is used for open hub service, and you can create tranformation/DTP for the destination. and in some case the DTP delta is enabled.
    my questions are
    1) in the context menu of open hub destination, I did not find something like manage, which could list all the DTP request executed for this open hub destination..but for normal infocube or ods, you can find a list.
    2) if the DTP could enable the delta, where can I delete the latest delta request, and have a new load..
    3) I assume the DTP could be excuted with batch integrated with process chain by process type 'DTP', right ? if the destination type is third party tool, can it also be integrated with process chain ?
    4) can give some obvious difference of InfoSpoke and open hub destination in terms of advantage and disadvantage..
    Thanks a lot for the feedback.
    Best Regards,
    Bin

    Hi ,
    Please look at the links below . hope it helps .
    http://www.sdn.sap.com/irj/scn/go/portal/prtroot/docs/library/uuid/d063282b-937d-2b10-d9ae-9f93a7932403?QuickLink=index&overridelayout=true
    http://www.sdn.sap.com/irj/scn/go/portal/prtroot/docs/library/uuid/e830a690-0201-0010-ac86-9689620a8bc9?QuickLink=index&overridelayout=true
    Regards
    Santosh

  • Issue for DTP from DSO to open hub destination

    Hello Gurus,
            I have a issue for DTP from DSO to open hub destination, long text for error in the monitor is as follows:
              " Could not open file
    SAPBITFS\tst_bi\bit\BIWork\GENLGR_OneTimeVendr_2 on application server"
              " Error while updating to target ZFIGLH03 (type Open Hub Destination)     "
          for open hub destination, I check the configure for logical file name , which is "
    SAPBITFS\tst_bi\bit\BIWork\GENLGR_OneTimeVendr",
    I am wondering where that file " 
    SAPBITFS\tst_bi\bit\BIWork\GENLGR_OneTimeVendr_2" in the error message comes from?
    Many thanks,

    Hi
    You do not need to create a file in application server. It will be created automatically.
    But if you have defined a logical file name in tcode FILE and used that in OHD and if it is not correct then it will show a conflict. Check this out.

  • Inconsistency with ZPMSENDSTATUS for Open Hub Destination

    Hi Experts,
    I dont know you have come accross this scenario but I will apprecaite all your advice.
    We have APO 7.0 and we are using the BW component to extract data to informatica.
    Our environment is as follow
    SAP SCM 7.0 SP8
    Informatica PWCD 8.1.6 hot fix 13
    1.we have created 5 chains with similar variants for ZPMSENDSTATUS
    2. Each variant relates to a folder and workflow to be executed within informatica
    3. When we execute the chains only two work correctly all the time and 3 fail
    the 3 that are failing are failing within the ZPMSENDSTATUS when running API RSB_API_OHS_DEST_SETPARAMS. This is odd since the other two only differ on the workflow to be executed and obviously on the Open Hub destination. Can anyone provide some advice on this?
    thanks,
    Nat

    Hi Maruthi,
    thanks for your response but here are the details for the variants that are working and failining
    working:
    DEST: < here we have the RFC that relates to the logical system and the program id which is also included in saprfc.ini>
    INFPARA: AERO/APO:WF_SAP_APO_PART_PLATFORM_ABP2_BGA:S_M_SAP_APO_PART_PLATFORM_ABP2
    CONTEXT: OHS API
    OHDEST: <OHDS defined as target>
    Failing:
    DEST: < here we have the RFC that relates to the logical system and the program id which is also included in saprfc.ini>
    INFPARA: AERO/APO:WF_SAP_APO_PART_PLATFORM_ABP4_ATR:S_M_SAP_APO_PART_PLATFORM_ABP4
    CONTEXT: OHS API
    OHDEST: <OHDS defined as target>
    So you see, they all target the same folder only different workflows. Still it works for the first and not for the second.

  • Multiprovider as Datasource for Open Hub Destination

    Hello,
    Can you pls let me know if we can used the Multiprovider as datasource for the Open Hub destination. Currently we are Bi7.0 and SP9. I am not able to see the option of Mutiprovider in the template.
    Appreciate the input ASAP.
    Thanks
    Gopal

    Multiprovider is not yet supported for Open Hub destination and hence you dont see that option.
    As of SAP NetWeaver 2004s SPS 6, the open hub destination has its own maintenance interface and can be connected to the data transfer process as an independent object. As a result, all data transfer process services for the open hub destination can be used.  You can now select an open hub destination as a target in a data transfer process.  In this way, the data is transformed as are all other BI objects.
    In addition to the InfoCube, InfoObject, and DataStore object, you can also use DataSources and InfoSources as templates for the field definitions of the open hub destination.
    The open hub destination now has its own tree under Modeling in the Data Warehousing Workbench. This tree is structured by InfoAreas.
    <b>Restrictions -</b>
    It is not currently possible to use MultiProviders as the source for the open hub destination.
    http://help.sap.com/saphelp_nw2004s/helpdata/en/43/7c836c907c7103e10000000a1553f7/content.htm
    Hope it Helps
    Chetan
    @CP..

  • Help required for Creation of process chain for Open hub destination(table)

    Hi Experts,
    I am working on creating process chain for the open hub destination which is targeting data transfer into Table. Which is then push to SAP PI.
    2.) Steps to create a process chain for OHD????? Not knowing and even after that
    3.) What are the steps for Pushing it to PI (I mean we have generated client proxy also) But need to know ABAP related to trigger process chain
    please need your expertise and knowledge. Well atleast with the point 2 I need BI experts help...
    Regards,
    [Gaurav Patwari|http://gauravpatwari.wordpress.com]

    Hi Gaurav,
    Creating Process chain for OHD.
    1. Go to RSPC-> Create a astart variant.
    2. Click on process typesin the menu. Fourth from the left. Under this go to LOAD PROCESSES AND POST PROCESSING option.
    3. The fifth option is "DATA EXPORT INTO EXTERNAL SYSTEMS". Click on that.
    4. Click on F4 and choose your infospoke name.
    5. Press the RIGHT button.(if it asks for variant, create a new variant in the same small window besideswritting bar).
    6. Join the START process and the 2nd process via dragging.
    7. Now add your required other processes.
    The general process chain processes are :
    1. Start.
    2. Delete Indexes (if data loaded to cube).
    3. Load process.
    4. Delete duplicate request (if it is a full load in cube).
    5. Create Indexes.
    Hope it helps.
    Preet

  • Can we set key figure in between 2 char in open hub destination.

    can any one tell me how can i insert a key description in between the char in open hub destinatio.
    Exp:my key description:quantity
    and char desc:material and plant.
    now i want field should be,
    Material.
    quantity
    plant
    when i am save the OHD ,i am getting the field like this,
    material
    plant
    quantity
    Can any one help me on this?

    Hi,
    If I am not wrong, u want to rearange the OHD fields
    thn
    select the field u want to insert and press the button "CUT"
    then place in the position and press "INSERT IN NEW ROW"
    Regards
    Md Zubair sha

  • Need a code for open hub destination

    Hi Experts,
    I have a requirement on open hub destination, as we are working on FI datasources,
    while using open hub destination, it will create two files. one is structure file and another is header file,
    nw my client is asking us to combine these both files into a same file while using open hub destination..
    could you please explain me hw to achieve this situation
    plz share the CODE ...
    Thanks in advance !!!!
    Regards
    sinu reddy

    ghhyh

  • How to export data from ODS through Open hub destination ?

    Hello Expert,
    I ahve some data in ODS.I want to export that data to some thirdt party application through Open Hub destination .How will I do that ??
    Eary reply is appreciated .

    HD IS :
             1.DATASOURCE
             2.INFOSOURCE
             3.DATASTORE OBJECT
             4.INFOCUBE
             5.INFOOBJECT
    1.Assume that info cube existing  with data
    2.Create openhuB services
        GO FOR OHD---->select Any  info area ->CONTEXTMENU>Create OHD
         Give the 'OBJECT NAME'----Continue
    In OHD Destination type Depends upon 3 types
       1.DB-TABLE
       2.FILE
       3.THIRD PARTY TOOL
      3.Go for the Defination Tab give the File name
    4.Create Transfermations
    5.Create DTP
    OPEN HUB SERVICES WITH -
    INFOSOURCE:OHD---DB TABLE SPECIFY:/BIC/OH
    1.Assume that info cube existing  with data
    2.Create openhuB services
        GO FOR OHD---->select Any  info area ->CONTEXTMENU>Create OHD
        Give the 'OBJECT NAME'----Continue
    3.Select Dbfile----->Continue
       Select file location where you want to store file .
        O.Semantic key
    4.Create Transfermations
    5.Create DTP
    Regards ,
    Praveenyagnamurthy.

  • Text Values in Open Hub Destination

    Hi ,
      Can anybody tell how to extract the text of particular info object using open hub destination into a file.
      For ex data is loaded from sales billing cube, ship to customer code is extracted , but need ship to customer name . how to extract this in open hub destination using BI 7.
    In queries , we use display key and text then get both key and text values. similar like that is it possible to get in open hub or not.
    Thanks .

    Hi Uppala,
    Following are 2 approaches to achieve this
    Approach 1:
    1. Create info spoke which will extract data from cube to data base table (do not use file)
    2. Create one more info spoke which will extract master data text to data base table (again do not use file)
    3. Create join on both the data base tables.
    4. Write a ABAP program which will create a file based on the joins of tables.
    Approach 2 :
    1. Create multiprovider on cube and master data info object
    2. Create info spoke based on the multi provider, that will extract data to flat file
    For 2nd approach u have to put certain filters to avoid unwanted rows
    Regards,
    Ajinkya

  • Use of Open Hub Destination to load data into BPC

    Hi Gurus,
    I want to load the data from SAP BI to BPC (NW version), using Open Hub Destination. I want to know few things about this.
    1) What method should I use? Should I use Destination as Flat Files or Database tables? If Database tables, how can I use this data in the tables in BPC?
    2) If I go for Flat Files, which  is saved in the application server of BI, how can I import those files to BPC and use it using Data Manager?
    3) Also, in case of Flat Files, there are two files which are created. One is the control file and other one is the data file. How can I use both of them? Or can I just use data file without header?
    Your replies will be much appreciated.
    Thanks,
    Abhishek

    Hi Anjali,
    I can use the standard data manager package from BI to BPC, if the CSV file is available in the BPC Server or in case if I am directly extracting from BW object, InfoObject or InfoCube.
    But, since I will be using Open Hub, the output of this can be in a Database Table or a CSV file, preferably in SAP Application Server. In such cases, how can I use the Database table or CSV file in the standard Data Manager Package?
    Thanks for your reply.
    Abhishek

  • Open Hub Destination question

    Hi,
    We want to be able to extract data from a cube and would like to use Open Hub Destination for it. We also want to extract deltas.
    The question is ..
    Should we be making a single info package for this with delta capability ..
    or 2 separate info packages .. 1 for full mode and another for delta mode.
    Regards
    Vandana

    Hi,
    Yes what you said it is write
    from which cube do u want to retract the daat that cube will be the source
    first create the open hub in RSBO tcode and
    in open hub destination select aht and create the transformation between the table( to where u want to export)
    then create the DTP between the cube to the table
    and directly u can use the delta DTP first all the records will come and later onle delta records will be updated.
    Thansk & Regards,
    sathish

  • Open Hub Destination error

    Hi,
    I am working on Open Hub (first time ofcourse). And performing following steps:-
    Modeling>Open Hub> InfoArea> Create Open Hub Destination> Output to File .csv. field names>DIrectory> activate> Create transformation> activate Transformations>Create DTP..Here I am getting an error that destination Open Hub is not active.
    Can anyone please tell me am i missing any step here?
    Thanks

    Hi...
    Always bear in mind that the OHD, once created, should be saved and activated, before creating any transformations or DTPs for it.
    And each time the OHD is changed, all the objects corresponding to it... (The transformations and the DTPs) should be activated.
    Hope this helps.
    Thanks
    Sree Ramya Lanka

Maybe you are looking for