Open Hub Destination Delta issue

Hi experts,
We have a scenario in which a Open hUb is created on top of a cube.
Regular delats are moving from this cube to Open hub daily.
But as per the request, we had to dump the cube data, and the reload took, 2 days to complete.
So, that 2 days, data got missed in Open hub.
Now, as the new request are not datamarted, next time wen the DTP between Cube and Open Hub runs, it will fetch all the requests right. we dont want that.
So, can we run a init w/o data transfer so that source status will be fetched?
but then how to load that 2 days data alone.
Kindly advise.
Edited by: neethacj on Mar 7, 2012 1:06 PM

If the cube doesnt have any other targets aprart from this OHD,
delete the two days requests in the cube, set the init from cube to ohd
load the cube for delta( brings two days requests) and run the ohd

Similar Messages

  • DTP Delta request for open hub destination.

    Hi expert,
    As you know, In BW 7.0, open hub destination is used for open hub service, and you can create tranformation/DTP for the destination. and in some case the DTP delta is enabled.
    my questions are
    1) in the context menu of open hub destination, I did not find something like manage, which could list all the DTP request executed for this open hub destination..but for normal infocube or ods, you can find a list.
    2) if the DTP could enable the delta, where can I delete the latest delta request, and have a new load..
    3) I assume the DTP could be excuted with batch integrated with process chain by process type 'DTP', right ? if the destination type is third party tool, can it also be integrated with process chain ?
    4) can give some obvious difference of InfoSpoke and open hub destination in terms of advantage and disadvantage..
    Thanks a lot for the feedback.
    Best Regards,
    Bin

    Hi ,
    Please look at the links below . hope it helps .
    http://www.sdn.sap.com/irj/scn/go/portal/prtroot/docs/library/uuid/d063282b-937d-2b10-d9ae-9f93a7932403?QuickLink=index&overridelayout=true
    http://www.sdn.sap.com/irj/scn/go/portal/prtroot/docs/library/uuid/e830a690-0201-0010-ac86-9689620a8bc9?QuickLink=index&overridelayout=true
    Regards
    Santosh

  • Open hub destination issue

    Hi,
    In our project we have Client Instance (eg: BWPD) and Application server (eg: BWPRD) defined for load balancing.
    We have created open hub and assigned destination server BWPD.
    When i execute the DTP manually in BWPD, it runs succesfully.
    However, the same DTP when placed in the process chain fails with error message :
    No Such File or Directory
    Could not open file D:\usr\sap\BWPD\A01\work\Material on application server
    Error while updating to target ZXXXX.
    Options Tried:
    Schedule process chain in the background server BWPD (same server which has been mentioned in Open hub dest) still DTP failed.
    Tried with Application server it failed.
    Tried with HOST as option it failed.
    couldn't make out what is going wrong. Any thoughts ?
    Regards.

    Hi there,
    found that doc quite useful.
    Maybe could shed some light to your issue.
    [Creating  Open Hub Destination  using a Logical file to extract the data|Creating  Open Hub Destination  using a Logical file to extract the data]
    Also, what OS do you have?
    is the Syntax Group accordingly created ?

  • Issue for DTP from DSO to open hub destination

    Hello Gurus,
            I have a issue for DTP from DSO to open hub destination, long text for error in the monitor is as follows:
              " Could not open file
    SAPBITFS\tst_bi\bit\BIWork\GENLGR_OneTimeVendr_2 on application server"
              " Error while updating to target ZFIGLH03 (type Open Hub Destination)     "
          for open hub destination, I check the configure for logical file name , which is "
    SAPBITFS\tst_bi\bit\BIWork\GENLGR_OneTimeVendr",
    I am wondering where that file " 
    SAPBITFS\tst_bi\bit\BIWork\GENLGR_OneTimeVendr_2" in the error message comes from?
    Many thanks,

    Hi
    You do not need to create a file in application server. It will be created automatically.
    But if you have defined a logical file name in tcode FILE and used that in OHD and if it is not correct then it will show a conflict. Check this out.

  • Open Hub Destination with Application Server Issue

    Hi all,
    I have requriment that i need to save my monthly data in application server with the help of OPen Hub Destination.I have created an DSO
    on that DSO i have created an Open HUb with application server as target.When i store my every month file in application server that file name should be a per the system date.
    So i have Created an LOgical File----ztest_data
    Assgined Physical path to Logical File-----/usr/sap/<sysid>........
    created an Logical File name------Ztest_data_file....
    IN the open HUb i have selected Aplication Server as My target and i have checked the box.
    I have selected Logical FIle name and given the file name which i have created-----ztest_data_file...
    Runned DTP it is showing no errors,but when i check in AL11 ,in the path which i have given(/usr/sap/<sysyid>........,i couldnt find the file.....
    Please check this and let me know...
    Thanks In Advance,
    Bobby.

    Hi,
    Please check the link /people/jyothi.velpula/blog/2010/01/20/creating-open-hub-destination-using-a-logical-file-to-extract-the-data .
    Also is your DTP full or delta. Try running the DTP as Full.
    Hope it helps.
    Best Regards,
    Kush Kashyap

  • Open hub destination of text and attribute for Functional Area

    Hi Expert
    I have a requirement to accommodate both attribute ( 0FUNCT_LOC_ATTR)  and text ( 0FUNCT_LOC_TEXT ) for functional area in open hub destination .For this a DSO has been created from both 0FUNCT_LOC_ATTR, 0FUNCT_LOC_TEXT data sources .Problem is that DSO has 0FUNCT_AREA ( Functional Area ) as only key fields . So text for both english ( EN ) and Dutch ( NL) for particular 0FUNCT_AREA not getting loaded .If we assign both 0LANGU and 0FUNCT_AREA as composite key of dso , data from attribute transformation is loaded as blank as there is no language filed in attribute data source ( 0FUNCT_LOC_ATTR ).   Please let me know how can I design to accommodate both text for different language and functional area for open hub destination  .
    I know that I can create separate open hub destination for master data TEXT and Attribute.
    Any help is highly appreciated.
    Regards
    Saikat

    HI Saikat,
    You can create an additional characteristics counter and make it a key field in the DSO.After that we can populate this characteristics in the end routine .It value would increase by 1 everytime  different text comes for the same functional area.
    Sample code
    data: counter type n,
             w_farea type /bio/oifunc_area.
    Sort result_package by func_area.
    count =0.
    *checking whether we have the same functional area or we have a different functional area.
    Loop at result_package assigning <Result_fields>.
    if w_farea is initial or w_farea = <result_fields>-func_area.
    count = count +1.
    else
    count =1.
    endif.
    w_farea= <result_fields>-func_area.
    <Result_fields>-zcount = count.
    ENDLOOP.
    Clear w_farea.
    This way the value of counter would get increased by one everytime a new text comes for the same functional area.We can have any number of texts for the same functional area by this method.
    The only issue would be when some delta update comes for the existing text in this method.If you are doing a full load to the DSO from the datasource, then there would be no issue with this method.

  • Open Hub Destination to Informatica

    HI All,
    I am having the following issue. We have created Open Hub Destination on InfoCube for Third Party (Informatica). But the target table is not visible in Informatica. They were able to view if created using InfoSpokes and able to extract using InfoSpokes. But when we use Open Hub Destination, they are not able to view them. We verified the RFC connection. Also triggered DTP for Open Hub Destination and it is sending data to the target teable structure.
    Please help me in this regard
    Regards
    YR

    Did you use Informatica's SAP Connector? It is a add-on and licese has to be purchased from Informatica. this will help to load delta and continuous daily load from SAP BI to target system with Informatica as ETL. InfoSpoke is a .csv file that created in a particuar folder thus Informatica able to see it.
    Thanks,
    S

  • Open Hub Destination question

    Hi,
    We want to be able to extract data from a cube and would like to use Open Hub Destination for it. We also want to extract deltas.
    The question is ..
    Should we be making a single info package for this with delta capability ..
    or 2 separate info packages .. 1 for full mode and another for delta mode.
    Regards
    Vandana

    Hi,
    Yes what you said it is write
    from which cube do u want to retract the daat that cube will be the source
    first create the open hub in RSBO tcode and
    in open hub destination select aht and create the transformation between the table( to where u want to export)
    then create the DTP between the cube to the table
    and directly u can use the delta DTP first all the records will come and later onle delta records will be updated.
    Thansk & Regards,
    sathish

  • Open Hub Objects Tranport Issue

    Hi,
    Environment - BI 7, SP level - 17.
    Trying to transport Open Hub destination DTP object (DSO to Flat File) from the Dev to QA box. Output file name is defined (Created Logical file) using Tcode - FILE. The Logical File Name is available in Dev, QA.
    In Dev box, we are able to write Output to a FLAT File (Application Server).
    While transporting DTP getting Transport error - Return Code - 12. (No clear message for this Error Message or Dump file created).
    Also applied OSS Note - 1175391 and 1241173.
    To overcome this issue, your valuable suggestions highly appreciated.
    Thanks,
    Madhu.

    Hi Madhu,
    I think the below link will helps you...
    https://www.sdn.sap.com/irj/sdn/go/portal/prtroot/docs/library/uuid/d063282b-937d-2b10-d9ae-9f93a7932403
    Regards
    Sudheer

  • Open hub destination with datasource as the source

    Hi,
    I am trying to use the new features of BI for open hub destination where you can use datasource as the source for extraction.
    I have created an infopackage, loaded data to PSA and from there to open hub destination via the DTP.
    When I try to execute the DTP, I get the following error:
    <b>Datapackage processing terminated.
    Error in substep
    You are not authorized to use this transformation.</b>
    I have been able to use the DTP to load open hub destination targets with datastore and infocubes as source.
    Also when we turn on the security trace, it doesn't fail anywhere.
    Has anybody faced a similar problem?
    Any help appreciated,
    Thanks,
    Payal.

    I have created to OHD - one for a 3.5 data source and one for a 7.0 data source (both are different data sources). While creating DTP for the 7.0 data source, it did create a default mapping for transformation.
    But I still get same error in both cases.
    When I create the DTP, it tries to create a transformation. But I get an information message :
    Exception CX_RS_MSG occurred (program: CL_RSBK_PATH==================CP, include: CL_RSBK_PATH==================CM00I, line: 16).
    I can still activate the DTP and when I run it says I don't have authorization to this transformation.
    I think there are issues while creating the transformation, but I don't know what!
    -Payal.

  • Open Hub Destination to Oracle database

    Hello All ,
    I am supposed to send data from my SAP BI (7.0) system to and Oracle Database using open hub destination. I have to use the option " Third Party " as destination while creating the Open Hub Destination .
    From my reading and acc to Basis I  realised that I need a 3rd party ETL tool in between BI and Oracle system to write data into the Oracle system since oracle is not SAP RFC aware . The 3rd party ETL that I am supposed to use in this is Business Objects - Data Integrator(BO-DI) . If anyone has implemented this type of Open hub service please throw  some light on this issue as to how to setup the 3rd party ETL in between and how to proceed .  Any how to document would also be useful .
    Thanks and Regards,
    Riddhi

    Hi,
    Pls go thru this links,
    http://help.sap.com/saphelp_nw70/helpdata/EN/43/79f902dfb06fc9e10000000a1553f6/frameset.htm
    http://help.sap.com/saphelp_nw70/helpdata/EN/43/79f902dfb06fc9e10000000a1553f6/frameset.htm
    Hope it helps u,
    Thanks & Regards,
    SD

  • Open Hub Destination over a Datasource as a source

    Hi Expert,
                    I am trying to create Open hub destination over Datasource as a source
    it showing me an error "Cannot connect DataSource to an open hub destination".
    Can anyone tell me is it possible to create an open hub destination over a datasource
    If yes then how?
    Thanks and Regards
    Lalit Kumar

    HI lalit,
    Check the below forum for refer they have refer using datasource as source for OHD m nt sure:
    Open hub destination with datasource as the source
    OHD ISSUE
    Thanks,
    Deepak
    Edited by: Deepak Machal on Dec 15, 2011 11:04 AM

  • Open hub destination ---Additional property needed--

    Hi,
    I have one open hub destination object which is writing data from infocube to an external file (in appliaction server ) and the file is distributed to customers via Email---
    This is absolutely working fine.
    Now, i also need to extract the data from other datasource to the same cube ( which are nothing but corrections to original data existing in cube ) and the corrected data needs to be updated to the file in applucation server!!
    as overwriting is not possible in infocubes i wonder how i could make the corrected data to be updated to the external files

    Hi,
    Try to load DELTA to the cube from the datasources. By this only the changes will be added to the already existing records.
    Else load the data from the cube to a new DSO and the load the data from other datasources to the same DSO. The latest correct data from other datasources will over-erite the existing data in the DSO.
    Use this new DSO in the openHub.
    Regards,
    Balaji V

  • Open Hub Destination with Flat File - Structure file is not correct

    Using the new concept for Open Hub destination (NOT an InfoSpoke).
    Output is to a Flat File. Remember that when output is to a flat file 2 files are generated: the output file and a structure file (S_<name of output file>.CSV)
    The Output file structure corresponds to the structure defined in the Open Hub destination.
    However, the S_ (structure) file does not reflect the proper structure. The S_ file lists all fields sorted in alphabetical order based on the field name, instead of listing all fields in the correct order that they appear in the output file.
    I have not found any settings that I can modify for the S_ file. When using InfoSpokes the S_ file is correct and I do not have to do anything.
    The issue is that the S_ file cannot be used to describe the output file. This is a problem.
    Anyone knows a solution to this issue? The goal is to get the S_ file to reflect the proper structure.
    I could not find any Notes on this. Please do not send general help links, this is a specific question.
    Thank you all for your support.

    Hi,
    However, the S_ (structure) file does not reflect the proper structure. The S_ file lists all fields sorted in alphabetical order based on the field name, instead of listing all fields in the correct order that they appear in the output file.
    As I know and checked this is not the case. The S_ file has the fields in the order of infospoke object sequence ( and in transformation tab, source structure).
    I would suggest you to control it again.
    Derya

  • Open Hub Destination Table Vs File

    Hello Experts,
    We have a requirement to send data from BW to 3rd party system every day. In the Open hub destination we have two options for the target.
    1. Table
    2. File
    ETL tool like Informatica will extract the data from table or file into 3rd party sytem.
    Which way is the best in performance and maintenance?
    Thanks in advance
    Sree

    Hi,
    If you follow the delta mechanism then table is good. You can also go for File not a problem. See the both eamples in the following articles.
    If you use Fiels then go for Application Server option, i.e. AL11 becasue no one can change teh files in this location, you can restrict the authorizations.
    Open Hub Destination: Part 1
    http://www.sdn.sap.com/irj/servlet/prt/portal/prtroot/com.sap.km.cm.docs/library/business-intelligence/m-o/open%20hub%20destination%3a%20part%201.pdf
    Open Hub Destination: Part 2
    http://www.sdn.sap.com/irj/servlet/prt/portal/prtroot/com.sap.km.cm.docs/library/business-intelligence/m-o/open%20hub%20destination%3a%20part%202.pdf
    Thanks
    Reddy

Maybe you are looking for