Open Hub with DataSource as Source of Transformation

Hello,
I have created an Open Hub Destination to a database table and activated it.  I now try to create a transformation with the new open hub as the destination and a custom datasource as the source of the transformation.
I get an error message that says "Cannot connect DataSource to an open hub destination".  We are on BI 2004s service pack 12.  I was able to perform this task successfully on previous support packs.
For instance, I cannot open an existing transformation from a datasource to the open hub that was created before the application of the support packs.  I get the same error message.
Any suggestions?
Thanks for your help,
Jason

Hi Michael,
I was loading directly from the DataSource into a database table because I was only going to use the table as a lookup table in a transformation.  The data doesn't really need to be in a BI object.
I did the same procedure one other time while we were on service pack 6, and it worked fine.
Anyway, I have loaded the lookup data into a DSO and everything is fine.  Thanks for your helpful response.
Regards,
Jason

Similar Messages

  • Archiving object in open hub with logical file name

    Hello,
    I am trying to use an open hub with a logical file name.
    By the SAP help looks like you have to use/define a archiving object:
    http://help.sap.com/saphelp_nw70/helpdata/en/e3/e60138fede083de10000009b38f8cf/frameset.htm
    http://help.sap.com/saphelp_nw70/helpdata/en/8d/3e4ec2462a11d189000000e8323d3a/frameset.htm
    Steps 1 and 2 of "Defining Logical Path and File Names" are done and tested:
           1.      Definition of the logical path name
           2.      Definition of the logical file name
           3.      Assignment of the logical fine name for the archiving object
    But is it not clear to me if an archiving object is suppose to be used, which one then?,  or if you have to create a new archiving object, with which characteristics?
    Any help welcome,
    Alberto.

    Alberto,
    Can you explain what you are trying to do. Are you trying to archive data from a BI object or are you trying to export data out of a BI object using open hub.

  • Open Hub - Data from multiple sources to single target

    Hello,
    I am using Open Hub to send the data from SAP BI to Flat files. I have a requirement where I want to create a single destination from multiple sources. In other words in BI we have different tables for attributes and text. I would like to combine the data from attributes  and text into a single file. For eg. I want to have material attributes and text in the same single file as output.
    Is this possible in Open Hub? If yes could you please help me to understand the process.
    Thanks,
    KK

    Hi,
    1. Create the Info Spoke and activate it
    2. Change it and go to transformation
    3. Check the box InfoSpoke with Transf. Using BAdI
    4. You are asked if you want to generate the spoke. Say yes & Simply set some texts and activate here, then return.
    5. You can now change the target structure. Simply create a Z structure with all the attributs & text field in it in the SE11 and enter it here.
    6. Double click on BAdI implementation & then double click again on "TRANSFORM" method of the implementation. It will take you to method  
    "IF_EX_OPENHUB_TRANSFORM~TRANSFORM"
    7. Write a code to select & fill the text field & map other filed with the attribute fields.
    Example:
    ZEMPLOYEE_ATTR_STRU - Target Structure for InfoSpoke EMPLOYEE_ATTR
    EMPLOYEE     /BI0/OIEMPLOYEE     NUMC     8     0     Employee
    DATETO     /BI0/OIDATETO     DATS     8     0     Valid to
    DATEFROM     /BI0/OIDATEFROM     DATS     8     0     Valid from
    COMP_CODE     /BI0/OICOMP_CODE     CHAR     4     0     Company code
    CO_MST_AR     /BI0/OICO_MST_AR     CHAR     4     0     Controlling Area of Master Cost Center
    HRPOSITION     /BI0/OIHRPOSITION     NUMC     8     0     Position
    MAST_CCTR     /BI0/OIMAST_CCTR     CHAR     10     0     Master Cost Center
    TXTMD     RSTXTMD     CHAR     40     0     Medium description
    Note: Text and attribute are in same structure.

  • Open hub service - Cube as source system

    Dear gurus,
    i am new to open hub concept.
    i need to transfer values from my query to a custom table in R/3.
    im using bi7.
    Questions:
    1) i guess pen hub is my solution. is it correct?
    2) do i have to transfer values from my cube or can i send values from my query too?
    3) i have created openhub service under modelling tab, i have selected target type as TAB. it says database table /BIC/OHZTABLO.
    is it correct?
    4) then i am trying to create Transformation from open hub ZTABLO to/ BIC/OHZTABLO. ?
    in that case it shows that transformation target is DEST- ZTABLO? i mean open hub service is my destination? then what is the source?i am confused in that point.
    5) Thanks all!

    1) i guess pen hub is my solution. is it correct?
    ->If u want to transfer dat from yr query then APD (analysis Process designer) will be the best bet to go for. Create APD and schedule it in process chain
    2) do i have to transfer values from my cube or can i send values from my query too?
    ->Using OH(open hub) you have to use Cube as a source,,,you cant use query then.
    3) i have created openhub service under modelling tab, i have selected target type as TAB. it says database table /BIC/OHZTABLO.
    ->ya,correct,there are two ways to do that,,,either u send the cube data to a open hub table(as its happening in yr case) or send it to application server/flat file
    is it correct?
    --> yes
    4) then i am trying to create Transformation from open hub ZTABLO to/ BIC/OHZTABLO. ?
    in that case it shows that transformation target is DEST- ZTABLO? i mean open hub service is my destination? then what is the source?i am confused in that point.
    ---> yr source is InfoCube and target is Open hub table /BIC/OHZTABLO
    5) Thanks all!
    Than you.

  • Open Hub with Attributes and Texts

    Hi,
    I´d like to know if it´s possible to have one open hub (database table) with texts and attributes.
    There´s no problem when I create it or it´s transformations and DTP´s, but when I try to load texts, after attributes load, it always generates a dump: SAPSQL_ARRAY_INSERT_DUPREC.
    I´d like to know a way without using a DSO, any suggestions?!
    Kind Regards,
    Tomas

    Hi,
    Pls chk SAP notes system, check SAP note 668466 for some help for this error.
    *As according to SAP NOTE ;668466*
    Summary
    Symptom
    You load master data attributes or texts for a characteristic. The loading terminates with runtime errors of the type SAPSQL_ARRAY_INSERT_DUPREC or with error message RSDMD 199. The termination point is in the DBFIRE method of the CL_RSDMD_UPDATE_MASTER_DATA class.
    Other terms
    Loading master data, attributes, texts, SAPSQL_ARRAY_INSERT_DUPREC, CL_RSDMD_UPDATE_MASTER_DATA, DBFIRE
    Reason and Prerequisites
    When you load attributes or texts, data records are usually inserted in the master data tables of the characteristic using an array insert statement (P, Q, X, Y, T tables). This results in a SAPSQL_ARRAY_INSERT_DUPREC runtime error if data records that are to be inserted and that relate to the unique primary key of the relevant database table already exist in this table. The following possible reasons are currently known for the occurrence of the SAPSQL_ARRAY_INSERT_DUPREC runtime error:
    1. Data inconsistencies in the master data tables of the characteristic
    SAPSQL_ARRAY_INSERT_DUPREC runtime errors or RSDMD 199 may occur if the data appears in an inconsistent status in the master data tables of the characteristic before you load attributes or texts. Notes 323140, 566044, 592757 and 599269 describe this (among other things) in more detail.
    2. Parallel loading of nondisjunct data packages
    If master data attributes are loaded over several dialog processes that are running in parallel, you must guarantee that the data packages of the relevant request are strictly disjunct concerning the characteristic value. This means that all data records that belong to a certain characteristic value may only be contained in a single data package. If data records of a certain characteristic value are distributed over several data packages, this may cause conflicts during the parallel loading of these packages when the master data tables are accessed and this may result in SAPSQL_ARRAY_INSERT_DUPREC runtime errors.
    The same conditions that apply to loading attributes also apply to loading language-independent master data texts. If language-dependent texts are loaded in parallel, the data packages must be disjunct in relation to the combination of characteristic value and language indicator.
    Note 566044 also describes the problematic nature of nondisjunct data packages or duplicate or overlapping data records.
    3. Parallel loading of time-dependent attributes or texts
    When you load time-dependent attributes or texts, a data record that is to be loaded is assigned to a characteristic value and also specifically to a time interval within which the contained attribute values or texts are valid. For this reason, requests for loading time-dependent attributes or texts typically contain several data records for each characteristic value. If data records of a certain characteristic value or the combination of characteristic value and language indicator are distributed over several data packages, this may (as already explained) cause conflicts when the master data tables are accessed and may result in SAPSQL_ARRAY_INSERT_DUPREC runtime errors or error message RSDMD 199.
    4. Delta update of an ODS object in a characteristic
    If you use a delta update from an ODS object to load master data attributes or texts, the activation of the ODS object determines the contents of the data packages that are used for loading the attributes or texts. If all of the ODS loading requests that are used as the data basis for the update into the characteristic are activated together in one step, this guarantees that the system sets up the data packages strictly disjunct in relation to the key fields of the ODS object. If the key fields of the ODS object are displayed one for one on the compounded characteristic value in the update rules, the data may be updated to the characteristic in parallel and by package.
    However, if ODS loading requests that were activated at different times appear in the data basis for the update, data records of a certain ODS key or characteristic value may be distributed over several data packages. As already explained, this may cause conflicts when you access master data tables of the characteristic, and it may result in SAPSQL_ARRAY_INSERT_DUPREC runtime errors or error message RSDMD 199.
    This problem is also described in Note 666213.
    5. Simultaneous loading of attributes and transaction data
    If, when loading attributes of a characteristic, you discover that there are still no entries in the master data tables (P, Q) for a specific characteristic value, the system prepares to insert corresponding data records into these database tables. You can load transaction data into an InfoCube or an ODS object, for example, with the option to create missing master data entries for characteristics that involved during the loading. If a request now creates entries for transaction data in the master data tables (P, Q) of the characteristic before the process has inserted corresponding data records into the master data tables (P, Q) for loading attributes for this characteristic, this may result in a SAPSQL_ARRAY_INSERT_DUPREC runtime error or error message RSDMD 199.
    This problem may also occur if you load attributes simultaneously for a characteristic that contains the other characteristic as a navigation attribute or compounding part. If a characteristic is compounded to other characteristics or if it contains characteristics other than the navigation attributes, the system also creates SID values for these dependent characteristics when the attributes are loaded depending on the settings in the InfoPackage.
    Solution
    To repair a possible inconsistency in the data, implement the RSRV test for master data.
    To avoid SAPSQL_ARRAY_INSERT_DUPREC runtime errors or error message RSDMD 199 when you load master data, carry out the following actions. The actions to be performed in each case depend on which of the above-mentioned reasons is responsible for the runtime error:
    1. Data inconsistencies in the master data tables of the characteristic
    If data inconsistencies are responsible for the runtime error, you can usually use the tests in transaction RSRV or the RSDMD_CHECKPRG_ALL program to solve this problem. Notes 323140, 566044, 592757 and 599269 describe these (among other things) in more detail. If you cannot implement an automatic correction in this way, you may have to use an ABAP program that is adjusted to the relevant situation to restore the data consistency.
    2. Parallel loading of nondisjunct data packages
    If the runtime error does not lead back to the parallel loading of nondisjunct data packages, you must identify the duplicate data records that cause the problem. As described in Note 566044, you can change the insertion into master data tables from one array insert to an individual insert statement. If you use a single record insert statement to insert data records, duplicate data records are logged in the monitor by error messages (RSDMD 196). If the data records of a failed request are still available in the PSA, you can also search for duplicate data records directly in the relevant PSA table.
    You may also create duplicate data records if key values are changed in the transmission rules or update rules. If relevant rules are defined, you should check these to see if they can possibly create duplicate data records. If this is the case, you must correct the rules accordingly.
    3. Parallel loading of time-dependent attributes or texts
    The extractors used for loading time-dependent attributes or texts cannot guarantee that all data records assigned to a certain key value are contained together in only one data package. For this reason, to avoid SAPSQL_ARRAY_INSERT_DUPREC runtime errors or error message RSDMD 199, you should not load time-dependent attributes and texts in parallel. Data packages are processed serially if the 'PSA only' and 'Update Subsequently in Data Targets' settings under 'Processing' are used in the InfoPackage.
    4. Delta update of an ODS object in a characteristic
    To avoid SAPSQL_ARRAY_INSERT_DUPREC runtime errors or error message RSDMD 199 during the delta update from an ODS object to a characteristic, you must ensure that the data basis used for the update to the characteristic consists exclusively of ODS loading requests that were activated together in one step only. Note 666213 describes in detail how you can ensure this.
    5. Simultaneous loading of attributes and transaction data
    If the runtime error occurs because another load task creates SID values or master data for the same characteristic at the same time when attributes are loaded, you can only solve the problem by rescheduling. The load task in question must be scheduled in such a way that you can exclude simultaneous processing.
    *pls assign points,if info is useful*
    Regards
    CSM reddy

  • Problem with open hub with application server

    Hello...
    I have a problem.
    I have bw 7.0 ad im usin Open Hub.
    When I configured the open hub in my development environment, In the destination tab i try two options
    Option 1.
    I use the application server, then i choose my server (the development server).
    But when i try to transport I get the follow mistake:
    Application sevier sapnwd01 does not exist (sapnwd01 is my development server)
    I now I can change the table  RSBFILE , or made an abap program.
    The problem is, i need to pass the transport withouts errors to quality, because i need to transport to my production server.
    And in production server I don't have the permission to change tables.
    Option 2.
    Using a logical file name (this logical file name, i used for other process and its working correctly)
    But when i try to transport I get the follow mistake:
    Unpermitted name   for a file .
    I hope you can help me!!
    Thanks in advance

    Hi Sharayu Kumatkar  
    Sorry for the delay, I had been problems with my access :$
    I check the AL11, there are a lot of directories, so I dont know which its the one
    My logicla path is ZAPD_INVERSIONES
    My phisical path is /netweaver/bw/fuentes/miig/altamira/
    My logical path Definition is :
    Logical file ZAPD_INVERSIONES2
    Physical file  ExtraccionInversiones.txt
    Data Format ASC
    Aplicat. Area FI
    Logical Path ZAPD_INVERSIONES
    In definition of variable
    FILENAME Test.txt
    I get this messaje error while Im trying to transport
    Start of the after-import method RS_DEST_AFTER_IMPORT for object type(s) DEST (Modo activación)
    Unpermitted name   for a file
    Thanks to help me!!

  • Open hub with tab separator

    Hello,
    I've an open hub to export to flat file. Is it possible to use a tab in the separator parameter?
    I'd like to have the exported data in diferent cells in my excel sheet. It should be enough with comma but all data appear in the same cell in the CSV.
    Can anybody help me?
    Thanks in advance.
    Regards.

    Hi,
    While exporting,  On the Destination tab page, select the required destination and in the separator option give ( ; ) instead of ( , )
    try this .I hope you will get your columns in diffrent cells in the excel sheet.
    please refere to the link also.
    http://www.sdn.sap.com/irj/scn/go/portal/prtroot/docs/library/uuid/501f0425-350f-2d10-bfba-a2280f288c59?QuickLink=index&overridelayout=true
    Regards,
    Sidhartha

  • Open Hub with step by step

    hi
    Can anybody tell me how a DTP can be used to transfer data from BI system to outside the BI system using Open hub. I need step by step.
    Regards,
    Wishva

    Hi,
    Go to ---> tcode RSBO,
              --->In the main menu of the Administrator Workbench, choose:
                   Tools _ Open Hub Service
                   Enter a technical name for your InfoSpoke.
              --->Give your target, 
              --->give the source and execute.

  • Transformations on Datasource for Open Hub Services

    Dear all,
         I have created a Generic Datasource on a transparent table. Following which  have created an Open hub service on this datasource. Now am unable to create the thansformations on this. I encountered the following error "Cannot connect DataSource to an open hub destination ". 
    Please let me know whether it is possible to generate Transformations for Open hubs on Datasource.
    If not, let me know any procedure to populate data from generic datasource.
    Kind Regards,
    Sunitha

    Hi,
    connection with an datasource is not possible. But, try to connect the datasource to with an infosource and use the infosource as a source for the open hub. maybe this works.
    cheers
    Juergen

  • Open Hub Destination with Application Server Issue

    Hi all,
    I have requriment that i need to save my monthly data in application server with the help of OPen Hub Destination.I have created an DSO
    on that DSO i have created an Open HUb with application server as target.When i store my every month file in application server that file name should be a per the system date.
    So i have Created an LOgical File----ztest_data
    Assgined Physical path to Logical File-----/usr/sap/<sysid>........
    created an Logical File name------Ztest_data_file....
    IN the open HUb i have selected Aplication Server as My target and i have checked the box.
    I have selected Logical FIle name and given the file name which i have created-----ztest_data_file...
    Runned DTP it is showing no errors,but when i check in AL11 ,in the path which i have given(/usr/sap/<sysyid>........,i couldnt find the file.....
    Please check this and let me know...
    Thanks In Advance,
    Bobby.

    Hi,
    Please check the link /people/jyothi.velpula/blog/2010/01/20/creating-open-hub-destination-using-a-logical-file-to-extract-the-data .
    Also is your DTP full or delta. Try running the DTP as Full.
    Hope it helps.
    Best Regards,
    Kush Kashyap

  • Data selection in new Open Hub destination Vs Infospoke

    In BW 3.5, we can use T code RSBO to create Infospoke, In the selection tab, we can use any infoobject in the source infocube as the data filterfor the infoSpoke, e.g I can set company code as 1002 etc, while in the new  BI 7.0, Open Hub Destination,  where I can achieve the same by set the filter for a specific infoobject in the source cude?
    Edited by: Jianbai on Mar 29, 2010 11:49 PM
    Edited by: Jianbai on Mar 29, 2010 11:51 PM

    For Open Hub in BI7 we are creating transformation and DTP. The same resrtiction you can give in DTP (activate it after change).

  • Open hub problem

    Hi Gurus,
    please help me out of this problem
    When we execute the dtp of open hub, it is creating a extract file on
    server directory. Our problem is record length of output record is
    always truncated to 110 characters. Record length of open hub structure
    we have crated is 151 charadters(excluding delimiter).
    We are not using infospokes as they are obsolate in 7.0.
    We are using NW BI 7.0 We are on Service pack 20
    and the excel 2000.
    Thank you

    Hi,
    Please let me know what is the length of the structure in the Open Hub with the delimiters.
    Btw,
    You can create a custom field in the open hub maintenance and specify its size to whatever size you desire with the delimiters and then assign all the source fields to this through a routine.
    Regards,
    Saikat

  • Open Hub fetching data from Change log table

    Hi Gurus,
    I have a scenario where in I have an Open Hub destination whose source is a DSO, and we are sending data into SAP Data services. We found that the data being sent is from the change log table. Is this how an Open Hub with DSO as a source works ? Or is there any possible way we can send data from the Active table of the DSO, when the DTP is run.
    Kindly help. Thanks in advance !
    Regards
    Snehith

    Hi,
    For Extraction from the DataStore Object, we have the following options for Delta init Extraction:
    Active Table (with Archive)
    The data is read from the active table and from the archive or from a near-line storage if one exists. You can choose this option even if there is no active data archiving process yet for the DataStore object.
    Active Table (Without Archive)
    The data is only read from the active table. If there is data in the archive or in a near-line storage at the time of extraction, this data is not extracted.
    Change Log
    The data is read from the change log of the DataStore object
    We can select any of these options for data trasnfer.
    Regards,
    Geetanjali

  • Data Extraction in Open Hub Destination using Process Chain

    Hi
    I want to extract data in Open Hub Destination (database table)from Data Store Object(ODS) through Process Chain.
    When i tried to create process chain, i found only one option under Process types - Data export into External System. Here it asks for Infospoke instead of Open Hub Destination.

    Michael is correct and below is the rationale...
    <a href="http://help.sap.com/saphelp_nw2004s/helpdata/en/43/58e1cdbed430d9e10000000a11466f/content.htm">http://help.sap.com/saphelp_nw2004s/helpdata/en/43/58e1cdbed430d9e10000000a11466f/content.htm</a>
    <b>Integration</b>
    You can use the data transfer process to update data to the open hub destination. The data is transformed in this process. Not all rules types are available in the transformation for an open hub destination: Formulas, reading master data, time conversion, currency translation and unit conversion are not available.
    Hope it Helps
    Chetan
    @CP..

  • Proses open hub services and proses chain.

    Re: OPEN HUB with Process Chain  
    Posted: Jul 28, 2007 10:29 AM    in response to: Viswanadh Vadde       Reply      E-mail this post 
    Dear Viswa,
    Please Forward me the Screen Shots of Sample Chain and data Flow]
    Thanks well in advance,
    ramireddy
    id:  [email protected]

    Re: OPEN HUB with Process Chain  
    Posted: Jul 28, 2007 10:29 AM    in response to: Viswanadh Vadde       Reply      E-mail this post 
    Dear Viswa,
    Please Forward me the Screen Shots of Sample Chain and data Flow]
    Thanks well in advance,
    ramireddy
    id:  [email protected]

Maybe you are looking for

  • What's a good web design software to use with hype.

    I didn't have luck with sandvox, rapid weaver, or highway. It seems that iWeb is gone and im trying to transition into a diffrent html5 compatible web software. I would like to learn web coding but for now i need a quick and efficiant transsition. I'

  • Check printing by amount authorization

    Hi, I have the following issue: There are two treasury areas in the company and one of then is the main one. One prints checks up to 10000 (local currency) if the amount is greater than 10000 the check must be printed in the main treasury. Is there a

  • How to delete the line between the last point and first point?

    How to delete the line between the last point and first point?  I want to draw a curve many times, from first point to the end point. and redraw from first point to the end point.But I hope update point by point. but between the end point and the fir

  • Application Language Problem

    Hello, I am writing From Turkey. I have an iphone 5s and i believe i am the one of first iphone 5s buyer in Turkey. I had a problem and i called apple and nobody solved my problem. I tunes help desk or whatever thay said that problem is not their pro

  • N80 Sms Reminder

    Hi all Can someone point me in the right direction, im after an sms reminder for my N80, dont mind paying as long as it works, can anyone help???? Thanks Pete ;-)