Open Hub  - Infosets as source and 'template fields' button

Hi all,
Two questions re Open Hub on 04s (we are on SP8).  I've surfed OSS, help.sap.com and SDN with no luck re answers to these:
- I've defined an infoset as my source of an open hub destination.  It appears I must create a transformation, but it generates no proposal, and I'm not sure how/if I can add a virtual object (such as an infoset) as a source in a transformation.  It appears that the DTP requires this transformation to be defined.  Am I missing something?  So far, testing on all other objects has been successful.
- How does one use the button 'Template Fields' when defining fields in an OHD?  I expected to see options to insert groups of fields (ex. all fields of an infoobject and its attributes), but I don't see any options available.
Thanks!
Lonnie

Update:
I was able to create the transformation (issue found btw keyboard and chair).
New problem:  I am unable to activate my DTP.  When choosing the source for my DTP, I get a popup window (as normal) to choose the possible sources, but the infoset is not available for selection.  Only the 'pieces' of the infoset (the two DSO's that comprise the infoset), and this doesn't help me because my source is the infoset, not the indivudual tables of it. 
  Also, I cannot change the extraction mode to FULL.  Only delta is available.
Thanks in advance for any help.

Similar Messages

  • Issue in mapping the same source and target fields

    Hi All,
    I am working on PI 7.0 and currently I am connecting to PI 7.0 via citrix desktop . Problem here is : in the mapping , direct mapping option that is available to map the same source and target fields is disabled . I do know that this is due to some java version issue .
    What I want to know is there any possibilty of solving this issue without installing an updated java version ?
    Regards
    Vinay P.

    I've faced the same for many times and all of that were due to Java version.
    1st of all, you have to check the authorizations you have. If you have enough (DEVELOPER_J2EE, CONTENT_ORGANIZER_J2EE, and so), and - obvious but very often forgotten - you have to assure you're trying to change it in the message mapping change mode On.
    Then, last shot should be downgrade your Java version.
    The 1.4.2 JRE version allow you to use not only the Mapp the Identical field but also Dependencies, Clear Mapping and some other tools/buttons. As mentioned above, if there's a fix in the new Java releases, I do not know (hope so) and would appreciate if you could let us know.
    Tks.

  • Open hub destination of text and attribute for Functional Area

    Hi Expert
    I have a requirement to accommodate both attribute ( 0FUNCT_LOC_ATTR)  and text ( 0FUNCT_LOC_TEXT ) for functional area in open hub destination .For this a DSO has been created from both 0FUNCT_LOC_ATTR, 0FUNCT_LOC_TEXT data sources .Problem is that DSO has 0FUNCT_AREA ( Functional Area ) as only key fields . So text for both english ( EN ) and Dutch ( NL) for particular 0FUNCT_AREA not getting loaded .If we assign both 0LANGU and 0FUNCT_AREA as composite key of dso , data from attribute transformation is loaded as blank as there is no language filed in attribute data source ( 0FUNCT_LOC_ATTR ).   Please let me know how can I design to accommodate both text for different language and functional area for open hub destination  .
    I know that I can create separate open hub destination for master data TEXT and Attribute.
    Any help is highly appreciated.
    Regards
    Saikat

    HI Saikat,
    You can create an additional characteristics counter and make it a key field in the DSO.After that we can populate this characteristics in the end routine .It value would increase by 1 everytime  different text comes for the same functional area.
    Sample code
    data: counter type n,
             w_farea type /bio/oifunc_area.
    Sort result_package by func_area.
    count =0.
    *checking whether we have the same functional area or we have a different functional area.
    Loop at result_package assigning <Result_fields>.
    if w_farea is initial or w_farea = <result_fields>-func_area.
    count = count +1.
    else
    count =1.
    endif.
    w_farea= <result_fields>-func_area.
    <Result_fields>-zcount = count.
    ENDLOOP.
    Clear w_farea.
    This way the value of counter would get increased by one everytime a new text comes for the same functional area.We can have any number of texts for the same functional area by this method.
    The only issue would be when some delta update comes for the existing text in this method.If you are doing a full load to the DSO from the datasource, then there would be no issue with this method.

  • Open hub,infoset

    Hi all,
             can any one tell what is open hub services and infoset.........is it used in projects ,in which type of scenario they are used.
    Thanks in Advance.
    jeevan.

    Hi Jeevan,
    Please before posting a new question...you can get lots of good information.

  • Access Violation when trying to read from Open Hub in BW

    Hi,
    Can someone help me out. I use BODS version 14.1.1.210. I have setup the connection to BW source and imported metadata for my OpenHub - no problem.
    Then created DTP and processchain to load Open hub data in BW.
    Made dataflow in BODS with Open Hub Tables as source and maintained processchain to be executed. When executing BODS job I get the following error:
    6800 10580 JOB 17-03-2014 13:10:34 Job <Test_OpenHub> is terminated due to error <170101>.
    6800 10580 SYS-170101 17-03-2014 13:10:34 |Session Test_OpenHub|Data flow Test Openhub
    6800 10580 SYS-170101 17-03-2014 13:10:34 System Exception <ACCESS_VIOLATION> occurred. Process dump option is off. Process is not dumped.
    6800 10580 SYS-170101 17-03-2014 13:10:34 Call stack:
    6800 10580 SYS-170101 17-03-2014 13:10:34 (null)
    6800 10580 SYS-170101 17-03-2014 13:10:34 Registers:
    6800 10580 SYS-170101 17-03-2014 13:10:34 RAX=00000000747E1648  RBX=0000000000209AB8  RCX=0000000000209AB8  RDX=0000000000000160  RSI=000000000231DAC0
    6800 10580 SYS-170101 17-03-2014 13:10:34 RDI=0000000008BEBAC0  RBP=0000000000000000  RSP=00000000002099C0  RIP=000000007477730A  FLG=0000000000010206
    6800 10580 SYS-170101 17-03-2014 13:10:34 R8=0000000000000000   R9=00000000FFFFFFFF  R10=0000000000620064  R11=0000000008C93600  R12=000000000033D390
    6800 10580 SYS-170101 17-03-2014 13:10:34 R13=0000000081D6FEC0  R14=000000000020A008  R15=000000000020A030
    6800 10580 SYS-170101 17-03-2014 13:10:34 Exception code: C0000005 ACCESS_VIOLATION
    6800 10580 SYS-170101 17-03-2014 13:10:34 Fault address:  000000007477730A 01:000000000002630A
    6800 10580 SYS-170101 17-03-2014 13:10:34 C:\Windows\WinSxS\amd64_microsoft.vc80.crt_1fc8b3b9a1e18e3b_8.0.50727.4940_none_88df89932faf0bf6\MSVCP80.dll
    6800 10580 SYS-170101 17-03-2014 13:10:34 ==========================================================
    6800 10580 SYS-170101 17-03-2014 13:10:34 Collect the following and send to Customer Support:
    6800 10580 SYS-170101 17-03-2014 13:10:34 1. Log files(error_*, monitor_*, trace_*) associated with this failed job.
    6800 10580 SYS-170101 17-03-2014 13:10:34 2. Exported ATL file of this failed job.
    6800 10580 SYS-170101 17-03-2014 13:10:34 3. DDL statements of tables referenced in this failed job.
    6800 10580 SYS-170101 17-03-2014 13:10:34 4. Data to populate the tables referenced in the failed job. If not possible, get the last few rows (or sample of them) when
    6800 10580 SYS-170101 17-03-2014 13:10:34 the job failed.
    6800 10580 SYS-170101 17-03-2014 13:10:34 5. Core dump, if any, generated from this failed job.
    6800 10580 SYS-170101 17-03-2014 13:10:34 ==========================================================
    Any good ideas?

    Hello Erik,
    not sure about it, but did you see this post some hours after yours?
    DS 4.1 job aborts with Access Violation
    Also, did you check ST22 to see if it generates any dump?
    And if you're able to check the BODS log files, it might me helpful. They're located at %LINK_DIR%\log directory. Check the names with the execution of the job.
    Hope it helps!
    Regards.
    Bruna Dupim

  • Open Hub with DataSource as Source of Transformation

    Hello,
    I have created an Open Hub Destination to a database table and activated it.  I now try to create a transformation with the new open hub as the destination and a custom datasource as the source of the transformation.
    I get an error message that says "Cannot connect DataSource to an open hub destination".  We are on BI 2004s service pack 12.  I was able to perform this task successfully on previous support packs.
    For instance, I cannot open an existing transformation from a datasource to the open hub that was created before the application of the support packs.  I get the same error message.
    Any suggestions?
    Thanks for your help,
    Jason

    Hi Michael,
    I was loading directly from the DataSource into a database table because I was only going to use the table as a lookup table in a transformation.  The data doesn't really need to be in a BI object.
    I did the same procedure one other time while we were on service pack 6, and it worked fine.
    Anyway, I have loaded the lookup data into a DSO and everything is fine.  Thanks for your helpful response.
    Regards,
    Jason

  • LSMW Source and Field Mapping

    Hi,
      I'm trying to use the LSMW.  What is the fastest way to maintain the source field and relationship?? Is it possible to define the source field and length in a text file and load it in the LSMW??
    Regards,
    Kit

    Hi Kit,
            u can create source fields in faster way.
    menupath
    In the 3rd step i.e., Maintain source fields
    source fields -> table maintenance
    here u can fill all the fields at a time.
    u can also maintain the fields in a file and copy and paste them in the Table maintenance .
    In 5th step i.e.,  Maintain Field Mapping and Conversion Rules
    Extras -> Auto field Mapping
    Now u get the auto field mapping setting pop-up.
    click ok.
    Now u Get the auto field mapping proposel pop-up.
    Here check Target field and Source field and click on Accept proposel push button for all the fields.
    with this the source and target field mapping are done automatically.
    reward if helpfull
    raam

  • OpenHub using Infoset as source

    Hi gurus,
    which is the best approach to use an infoset as a source of an OpenHub destination? Any experiences on this?
    Thanks in advance!

    Hi Inaki,
    You can create openhub destination based on Infoset, I dont' have any realtime exp. but I just read throug tech. documents,
    1) Create a Open hub destination, on the same screen it will ask you the template. This template will be used to select the fields for your openhub destination.
    2) Once you select all the fields and create open hub destination activate it and goto context menu from RSA1 to create transformation on this open Hub destination.
    3) Select the source of transformation as InfoSet, do the field mapping.
    4) Create DTP and you are ready to do data loading.
    Regards,
    Durgesh.

  • ABAP routine in Open hub transformation

    Hello Experts,
    My BW consultant created a Open Hub service. For creating a .csv file from a cube.
    For that he had a transformation defined between the source and destination fields.
    I was asked to write a routine to bring the negative sign to front for one of the fields.
    I wrote the following code.
    >    DATA V_OUT(25).
    >
    >   CLEAR V_OUT.
    >
    >    V_OUT = SOURCE_FIELDS-G_AVVWHC.
    >
    >      CALL FUNCTION 'CLOI_PUT_SIGN_IN_FRONT'
    >        CHANGING
    >          VALUE = V_OUT.
    >
    >      RESULT = V_OUT.
    there was no error in the syntax check.
    Now, when the open hub is created/ran without routine(direct assignment) it works fine.
    But when the above code is ran, it TIMED OUT!!!!
    I am unable to find any reason for that.
    Could anybody please provide some guidance on this.
    Regards,
    DNP

    Hi Jerry,
    Thanks for your input again. I checked the dump in ST22 and looks like nothing is related to Routine-
    >Information on where terminated
    >    Termination occurred in the ABAP program "CL_SQL_STATEMENT==============CP" -
    >     in "EXECUTE_QUERY".
    >    The main program was "RSAWBN_START ".
    >
    >    In the source code you have the termination point in line 19
    >    of the (Include) program "CL_SQL_STATEMENT==============CM004".
    SourceCde
    >METHOD execute_query .
    >*
    >* Execute the given query.
    >*
    >
    >  DATA:
    >    stmt_ref TYPE REF TO data,
    >    sql_code TYPE i,
    >    sql_msg  TYPE dbsqlmsg,
    >    c        TYPE cursor.
    >
    >
    >* check if a statement was provided by the caller
    >  IF statement IS NOT SUPPLIED.
    >    RAISE EXCEPTION TYPE cx_parameter_invalid
    >          EXPORTING parameter = 'STATEMENT'.
    >  ENDIF.
    >
    >>  CALL 'C_DB_FUNCTION' ID 'FUNCTION'    FIELD 'DB_SQL'                     
    >>                       ID 'FCODE'       FIELD c_fcode_prepare_and_open
    >>                       ID 'CONNAME'     FIELD me->con_ref->con_name
    >>                       ID 'CONDA'       FIELD me->con_ref->con_da
    >>                       ID 'STMT_STR'    FIELD statement
    >>                       ID 'HOLD_CURSOR' FIELD hold_cursor
    >>                       ID 'INVALS'      FIELD me->parameters->param_tab
    >>                       ID 'CURSOR'      FIELD c
    >>                       ID 'SQLCODE'     FIELD sql_code
    >>                       ID 'SQLMSG'      FIELD sql_msg.
    > IF sy-subrc <> 0.
    >  31 *   call error handling macro
    >  32     m_handle_error sy-subrc sql_code sql_msg.
    >  33   ENDIF.
    >  34
    >  35 * reset the parameters in order to make the statement object reusable
    >  36 * for the next execution of an DML statement or query immediately;
    >  37 * but the cursor C itself must not be freed because it will still be
    >  38 * accessed by the created result set object; the allocated cursor
    Regards,
    DNP

  • Open Hub Destination question

    Hi,
    We want to be able to extract data from a cube and would like to use Open Hub Destination for it. We also want to extract deltas.
    The question is ..
    Should we be making a single info package for this with delta capability ..
    or 2 separate info packages .. 1 for full mode and another for delta mode.
    Regards
    Vandana

    Hi,
    Yes what you said it is write
    from which cube do u want to retract the daat that cube will be the source
    first create the open hub in RSBO tcode and
    in open hub destination select aht and create the transformation between the table( to where u want to export)
    then create the DTP between the cube to the table
    and directly u can use the delta DTP first all the records will come and later onle delta records will be updated.
    Thansk & Regards,
    sathish

  • Open Hub Destination error

    Hi,
    I am working on Open Hub (first time ofcourse). And performing following steps:-
    Modeling>Open Hub> InfoArea> Create Open Hub Destination> Output to File .csv. field names>DIrectory> activate> Create transformation> activate Transformations>Create DTP..Here I am getting an error that destination Open Hub is not active.
    Can anyone please tell me am i missing any step here?
    Thanks

    Hi...
    Always bear in mind that the OHD, once created, should be saved and activated, before creating any transformations or DTPs for it.
    And each time the OHD is changed, all the objects corresponding to it... (The transformations and the DTPs) should be activated.
    Hope this helps.
    Thanks
    Sree Ramya Lanka

  • Open Hub  - Add header line to file on server.

    Hi All,
    How can I add header description to the file being generated in the server from the Open Hub. The open hub generates a fiule with data fields only, however i want a header line.
    Example
    IF my data fields are displayed as below,
    John M 12 2344
    Thomas M 17 8960
    Lin F 10 8467
    Jan M 12 4323
    I need to add, description as NAME above John, Thomas, Lin etc.
    as GENDER above M,M,F etc.
    as CODE above 12, 17,10 etc. and
    as EXTENTION above 2344, 8960,8467 etc
    Can anyone help.
    Points will be assigned

    Hi,
    i don't think this can be done even in BADI.
    what you can do is once the Infospoke run is complete and the file generated , run a custom report for this.
    in the custom report open a file dataset , write header and then close the file dataset.
    then again open the file dataset in append mode , writhe the file that the infospoke generated and close the file dataset.
    include this custom program after you infospoke run in the process chain.
    Regards,
    Neeraj.

  • Open Hub fetching data from Change log table

    Hi Gurus,
    I have a scenario where in I have an Open Hub destination whose source is a DSO, and we are sending data into SAP Data services. We found that the data being sent is from the change log table. Is this how an Open Hub with DSO as a source works ? Or is there any possible way we can send data from the Active table of the DSO, when the DTP is run.
    Kindly help. Thanks in advance !
    Regards
    Snehith

    Hi,
    For Extraction from the DataStore Object, we have the following options for Delta init Extraction:
    Active Table (with Archive)
    The data is read from the active table and from the archive or from a near-line storage if one exists. You can choose this option even if there is no active data archiving process yet for the DataStore object.
    Active Table (Without Archive)
    The data is only read from the active table. If there is data in the archive or in a near-line storage at the time of extraction, this data is not extracted.
    Change Log
    The data is read from the change log of the DataStore object
    We can select any of these options for data trasnfer.
    Regards,
    Geetanjali

  • Sample source code for fields mapping in expert routine

    Hi All
    Iam writing the expert routine from dso to cube for example I have two fields in dso FLD1,FLD2
    same fields in infocube also ,can any body provide me sample abap code to map source fields to target fields in expert routine,your help will be heighly appreciatble,it's an argent.
    regards
    eliaz

    Basic would be ;
    RESULT_FIELDS -xxx = <SOURCE_FIELDS> -xxx
    you have the source fields as source, and result fields for as the target. In between you can check some conditions as in other routines of transformation.
    BEGIN OF tys_SC_1, shows your source fields ( in your case DSO chars and key figures)
    BEGIN OF tys_TG_1, , shows your result fields ( in your case Cube characteristics)
    Hope this helps
    Derya

  • Open Hub - Change from File to Database table

    Hello Experts;
    I want to change the Open Hub Service that we create from File to Database table.
    We are using the Delta Update. However, after I change the Open Hub, Activate the transformation and DTP, when I run it, I don't have any value in the table.
    If I run a new delta into my cube, in the table I only get the data from delta.
    Can anyone knows how to upload all the requests that I have in my cube for the first time I run the delta DTP?
    Thanks and regards;
    Ricardo

    Create a new Full DTP and execute, so that you can dump all of the current contents of the InfoCube into the table and push/pull that table to the target prior to the next delta retraction.
    Edited by: Dennis Scoville on Nov 17, 2009 9:37 AM

Maybe you are looking for

  • Prevent update to document after status is changed

    Hi all, I configured a new status profile for a sales document ITEM and want to stop user from editing the item once the status is changed to 'Close'. I have tried with Business Transaction 'Edit' (forbidden is set) and 'No Change Allowed' (allowed i

  • Get the response instead of loading it in the browser

    HI. I would like to make a little program that would read the response of a request to a web page. I'm cant find how to do it. Everything i can foud call's the URL but loads th result in the initial browser window. Heres what I want : 1- A friend as

  • Call block limit

    I am nearing the call blocking limit on my Comcast phone. What is Comcast doing to expand or make limitless this feature. To protect the shrinking landline use i would think Comcast would address the issue of predaroy foreign sammers who laugh at FCC

  • Issue Regarding date

    Hi All, i have a cutome page in which i have a Date fileld along with some other fields and a button for saving the data. i have to save the data after validation. my problem is if i am not selecting the date through the calendar icon and entering th

  • ORACLE CC&B BARCODE READING ERROR

    Hello, On scanning ACCOUNT ID using barcode reader on account search screen only partial ID is displayed. The result is the same on other search screens when their corresponding ID's are scanned. This issue happens especially on the pop up search scr