Export Quanityfields with Open Hub

Hi,
i want to export an field (with type Quanity with fixed unit kg) with an Open Hub into a csv-file.
All other fields (type currency and characteristics) are written into the csv-file. Only the quantity fields are blank.
Do I have to use any conversion exit? Anything special in the Infoobject or transformation to care about?

Hi kevinhuang  ,
RFC connection.......... You can check in SM59 tcode...Double click on the the desired Source system
or
RSA1-->Source systems and then right click on Source system and Check
To test RFC connection...RSA1 ....Source system....... Right click on desired source system >> Connection Parameter.....Test connection
Pls go thru this links,
http://help.sap.com/saphelp_nw70/helpdata/EN/43/79f902dfb06fc9e10000000a1553f6/frameset.htm
BI Open Hub Destination and data transfer process setup
http://www.ibm.com/developerworks/data/library/techarticle/dm-0802li/index.html
Hope this helps U.
Regards
Chandra Sekhar
Edited by: Chandra  Sekhar T on Apr 21, 2009 3:36 PM

Similar Messages

  • Export data from BW to oralce with open hub service.

    Hi Dear all,
    When I tried to export data from BW to oralce with open hub service. There is no available RFC destination for this oracle system. That means i need to create a RFC destination for the oracle server.
    How can i achive this, in SM59? But U don't know much how to configure this. Is there any expert can help me there?
    Kevin

    Hi kevinhuang  ,
    RFC connection.......... You can check in SM59 tcode...Double click on the the desired Source system
    or
    RSA1-->Source systems and then right click on Source system and Check
    To test RFC connection...RSA1 ....Source system....... Right click on desired source system >> Connection Parameter.....Test connection
    Pls go thru this links,
    http://help.sap.com/saphelp_nw70/helpdata/EN/43/79f902dfb06fc9e10000000a1553f6/frameset.htm
    BI Open Hub Destination and data transfer process setup
    http://www.ibm.com/developerworks/data/library/techarticle/dm-0802li/index.html
    Hope this helps U.
    Regards
    Chandra Sekhar
    Edited by: Chandra  Sekhar T on Apr 21, 2009 3:36 PM

  • Open Hub: How-to doc "How to Extract data with Open Hub to a Logical File"

    Hi all,
    We are using open hub to download transaction files from infocubes to application server, and would like to have filename which is dynamic based period and year, i.e. period and year of the transaction data to be downloaded. 
    I understand we could use logical file for this purpose.  However we are not sure how to have the period and year to be dynamically derived in filename.
    I have read in sdn a number of posted messages on a similar topic and many have suggested a 'How-to' paper titled "How to Extract data with Open Hub to a Logical Filename".  However i could not seem to be able to get document from the link given. 
    Just wonder if anyone has the correct or latest link to the document, or would appreciate if you could share the document with all in sdn if you have a copy.
    Many thanks and best regards,
    Victoria

    Hi,
    After creating open hub press F1 in Application server file name text box from the help window there u Click on Maintain 'Client independent file names and file paths'  then u will be taken to the Implementation guide screen > click on Cross client maintanance of file name > create a logical file path by clicking on new entiries > after creating logical file path now go to Logical file name definition there give your Logical file , name , physical file (ur file name followed by month or year what ever is applicable (press f1 for more info)) , data format (ASC) , application area (BW) and logical path (choose from F4 selection which u have created first), now goto Assignment of  physical path to logical path > give syntax group >physical path is the path u gave at logical file name definition.
    however we have created a logical path file name to identify the file by sys date but ur requirement seems to be of dynamic date of tranaction data...may u can achieve this by creating a variable. U can see the help from F1 that would be of much help to u. All the above steps i have explained will help u create a dynamic logical file.
    hope this helps u to some extent.
    Regards

  • Extract Data with OPEN HUB to a Logical Filename

    Hi Experts,
    Can anybody help me in sending the link for How to guide...Extract Data with OPEN HUB to a Logical Filename?
    Thanks in advance.
    BWUser

    Hi,
    check this links...
    http://searchcrm.techtarget.com/generic/0,295582,sid21_gci1224995,00.html
    https://www.sdn.sap.com/irj/servlet/prt/portal/prtroot/docs/library/uuid/e698aa90-0201-0010-7982-b498e02af76b
    https://www.sdn.sap.com/irj/servlet/prt/portal/prtroot/docs/library/uuid/1570a990-0201-0010-1280-bcc9c10c99ee
    hope this may help you ..
    Regards,
    shikha

  • Need help with open hub

    Hi eveybody,
             I was trying to wok with open hub. Created infospoke, destination i gave was a csv file named d:/openmara.csv path, saved and activated the info spoke. I went to d:/ to view the uploaded files, two file was created, when i try to open a file it says "unable to read" , when i tried to open the second one i could open the file and see some datas but not in a correct format, why is that?
    Then i tried to load the data in database, so in the infospoke i selected the option for destination as database,saved and activated the infospoke, now where should i go to see the uploaded datas physically?
    Could anyone help me with this.
    Thanks,
    RR.

    Hi,
    Thanks for the reply. I created a infospoke and destination was database option. Opened the se16 gave the table name which starts with /bic/....(please correct me if i am wrong),then the initial screen of se16 came up with all the fields , but when i executed it, the table is empty.Could you say why it is happening.
    Thanks,
    RR.

  • Problem with open hub with application server

    Hello...
    I have a problem.
    I have bw 7.0 ad im usin Open Hub.
    When I configured the open hub in my development environment, In the destination tab i try two options
    Option 1.
    I use the application server, then i choose my server (the development server).
    But when i try to transport I get the follow mistake:
    Application sevier sapnwd01 does not exist (sapnwd01 is my development server)
    I now I can change the table  RSBFILE , or made an abap program.
    The problem is, i need to pass the transport withouts errors to quality, because i need to transport to my production server.
    And in production server I don't have the permission to change tables.
    Option 2.
    Using a logical file name (this logical file name, i used for other process and its working correctly)
    But when i try to transport I get the follow mistake:
    Unpermitted name   for a file .
    I hope you can help me!!
    Thanks in advance

    Hi Sharayu Kumatkar  
    Sorry for the delay, I had been problems with my access :$
    I check the AL11, there are a lot of directories, so I dont know which its the one
    My logicla path is ZAPD_INVERSIONES
    My phisical path is /netweaver/bw/fuentes/miig/altamira/
    My logical path Definition is :
    Logical file ZAPD_INVERSIONES2
    Physical file  ExtraccionInversiones.txt
    Data Format ASC
    Aplicat. Area FI
    Logical Path ZAPD_INVERSIONES
    In definition of variable
    FILENAME Test.txt
    I get this messaje error while Im trying to transport
    Start of the after-import method RS_DEST_AFTER_IMPORT for object type(s) DEST (Modo activación)
    Unpermitted name   for a file
    Thanks to help me!!

  • SAP BI 7.0 to SAP PI to FTP and with Open Hub Destination ...Help!!!!

    Dear SCN Experts,
    I am currently working on scenario where I have a requirement to push data from SAP BI 7.0 to SAP PI 7.
    And I am using Client proxy for SAP BI to send the data from BI tables to SAP PI and then to write to FTP address.
    Now the challenge I am facing is the ABAP Proxy how to use it with Process Chain and also I am having Open hub destination created for the same.(Specifically with the new version i.e. 7.0)
    Can you atleast make me understand what are the steps involved in this for Client Proxy, Process Chaing, How will proxy trigger and other regarding the same.
    I have searched SDN but got the document having older versions in it. Which doesn't serve the purpose.
    Regards,
    [Gaurav Patwari|http://gauravpatwari.wordpress.com]

    Hi Michal,
    Thanks for the reply buddy.
    I know that we can run the scheduled report for the proxy to fetch the report. But the client requirement is to use process chain, Open hub destination which fetches data from the ODS to Ztable created.
    We need to fetch that data from the table via our proxy. I am familiar with that report related method.
    I have one document using the same method but is using Infospoke(which is now obsolete) so have to OHD and also some of the proxy of XI's older version. So is not helping me out.
    Please do the needful. Or can you send me some sample scenario like this with screen shots. It will be a great help.
    Regards,
    [Gaurav Patwari|http://gauravpatwari.wordpress.com]

  • How to get rid of zeros for the data which has been extracted with open hub

    Hi all
    I am trying to extract data from the cube thru an open hub, but the data which i have extracted, for a cost center i am able to see '0's which is expected to be an empty space, even in cube there is an empty space.
    Can anyone suggest me quickly, points will be rewarded
    thanks
    preethi......

    In the transformation of the Open Hub,use BAPI....
    Create a target structure with all fields as type characters,move data from source structure to target structure in the BAPI using ABAP codes.When there is no value in character fields it would not show up 0's.

  • How to save a open request with open hub?

    Dear experts,
    we want to save a query result in an open hub destination. So at first we want to store the data in an ods object (we use bw 3.5) which was filled by the apd. The we want to tranfer the delta to the open hub destination which is connected with our bpc ms system. How can we care for extracting only closed requests from the open hub? And is it possible to extract the data from an ods which was filled by an apd??
    Thanks an regards,
    Gerrit

    Hi,
    In APD we use direct update DSO. SO its better to load the data from this DSO to one standard DSO or a cube. Then you can create a delta DTP, which can fetch request based delta to open hub destination.
    Hope it helps,
    Pravender

  • How to load delta, changing the name of file for every day with open hub?

    Hi gurus,
    I load daily delta data from ODS to flat file, using open hub.
    The file is located on server. Now, if I do delta load from ODS to flat file, every day the file is overwritten, because it has the same name.
    I want to load delta data everyday to different file name, for example BW_20060101 , BW_20060102, BW_2006_01_03 ....
    How do I do that?
    Best regards,
    Uros

    Hi Uros,
    This thread may give you some idea
    Changing the file name n Flat file Extraction
    https://websmp107.sap-ag.de/~form/sapnet?_FRAME=CONTAINER&_OBJECT=011000358700002201112003E
    Regards,
    BVC

  • Use transaction FILE to store data from a cube into a file with Open Hub

    Hi people:
    I'm using BI 7.0 .Mi requirement is to make a flat file using the information of a virtual cube. The file name must have the number of the month and the year. I know that this is possible through FILE transaction.
    Can anybody give me a clue how this transaction is used?Which are the steps in order to assemble the name of the file? Or is there any other option? I have defined the physical directory where the file must be leaved
    Any help will be great. Thanks in advanced

    Hi,
    pick up the code which you need from below.
    REPORT RSAN_WB_ROUTINE_TEMP_REPORT .
    TYPES: BEGIN OF y_source_fields ,
             /BIC/ZTO_ROUTE TYPE /BIC/OIZTO_ROUTE ,
             ZINT_HU__Z_WM_HU TYPE /BIC/OIZ_WM_HU ,
             CREATEDON TYPE /BI0/OICREATEDON ,
             ROUTE TYPE /BI0/OIROUTE ,
             PLANT TYPE /BI0/OIPLANT ,
             PLANT__0STREET TYPE /BI0/OISTREET ,
             PLANT__0CITY TYPE /BI0/OICITY ,
             PLANT__0REGION TYPE /BI0/OIREGION ,
             PLANT__0POSTAL_CD TYPE /BI0/OIPOSTAL_CD ,
             /BIC/ZRECVPLNT TYPE /BIC/OIZRECVPLNT ,
             ZRECVPLNT__0STREET TYPE /BI0/OISTREET ,
             ZRECVPLNT__0CITY TYPE /BI0/OICITY ,
             ZRECVPLNT__0REGION TYPE /BI0/OIREGION ,
             ZRECVPLNT__0POSTAL_CD TYPE /BI0/OIPOSTAL_CD ,
             KYF_0001 TYPE /BI0/OIDLV_QTY ,
             ROUTE__Z_CR_DOCK TYPE /BIC/OIZ_CR_DOCK ,
             REFER_DOC TYPE /BI0/OIREFER_DOC ,
           END OF y_source_fields .
    TYPES: yt_source_fields TYPE STANDARD TABLE OF y_source_fields .
    TYPES: BEGIN OF y_target_fields ,
             RECORDTYPE TYPE /BI0/OISTREET ,
             CONTAINER TYPE /BI0/OICITY ,
             /BIC/ZTO_ROUTE TYPE /BIC/OIZTO_ROUTE ,
             TRACKINGNUMBER TYPE /BIC/OIZ_WM_HU ,
             PO TYPE /BI0/OICITY ,
             STAGEDDATE TYPE /BI0/OICITY ,
             MOVEMENTTYPE TYPE /BI0/OICITY ,
             ROUTE TYPE /BI0/OIROUTE ,
             PLANT TYPE /BI0/OIPLANT ,
             PLANT__0STREET TYPE /BI0/OISTREET ,
             PLANT__0CITY TYPE /BI0/OICITY ,
             PLANT__0REGION TYPE /BI0/OIREGION ,
             PLANT__0POSTAL_CD TYPE /BI0/OIPOSTAL_CD ,
             ORIGINCONTACTNAME TYPE /BI0/OISTREET ,
             ORIGINCONTACTPHONE TYPE /BI0/OISTREET ,
             /BIC/ZRECVPLNT TYPE /BIC/OIZRECVPLNT ,
             ZRECVPLNT__0STREET TYPE /BI0/OISTREET ,
             ZRECVPLNT__0CITY TYPE /BI0/OISTREET ,
             ZRECVPLNT__0REGION TYPE /BI0/OISTREET ,
             ZRECVPLNT__0POSTAL_CD TYPE /BI0/OISTREET ,
             DESTINATIONCONTACTNAME TYPE /BI0/OISTREET ,
             DESTINATIONCONTACTPHONE TYPE /BI0/OISTREET ,
             RCCCODE TYPE /BI0/OISTREET ,
             GLCORCLLICODE TYPE /BI0/OISTREET ,
             JFCCODE TYPE /BI0/OISTREET ,
             DESCRIPTIONOFWORK1 TYPE /BI0/OISTREET ,
             DESCRIPTIONOFWORK2 TYPE /BI0/OISTREET ,
             INSTRUCTIONS TYPE /BI0/OISTREET ,
             REQUESTEDSHIPDATE TYPE /BI0/OICITY ,
             ROUTE__Z_CR_DOCK TYPE /BIC/OIZ_CR_DOCK ,
             REQUESTEDDELIVERYDATE TYPE /BI0/OICITY ,
             ATTSEORDER TYPE /BI0/OICITY ,
             CUBE TYPE /BI0/OISTREET ,
             WEIGHT TYPE /BI0/OISTREET ,
             PIECES TYPE /BI0/OIREFER_DOC ,
             REEL TYPE /BI0/OISTREET ,
             REELSIZE TYPE /BI0/OISTREET ,
             VENDORSKU TYPE /BI0/OISTREET ,
             ATTSESKU TYPE /BI0/OISTREET ,
             COMPANYNAME TYPE /BI0/OISTREET ,
             OEM TYPE /BI0/OISTREET ,
             REFER_DOC TYPE /BI0/OIREFER_DOC ,
             REFERENCENUMBER2 TYPE /BI0/OISTREET ,
             REFERENCENUMBER3 TYPE /BI0/OISTREET ,
             REFERENCENUMBER4 TYPE /BI0/OISTREET ,
           END OF y_target_fields .
    TYPES: yt_target_fields TYPE STANDARD TABLE OF y_target_fields .
    Begin of type definitions -
    *TYPES: ...
    End of type definitions -
    FORM compute_data_transformation
         USING     it_source TYPE yt_source_fields
                   ir_context TYPE REF TO if_rsan_rt_routine_context
         EXPORTING et_target TYPE yt_target_fields .
    Begin of transformation code -
      DATA: ls_source TYPE y_source_fields,
            ls_target TYPE y_target_fields,
            var1(10),
            var2(10),
            year(4),
            month(2),
            day(2),
            date(10),
            it_workdays type table of /bic/pzworkdays,
            wa_workdays type /bic/pzworkdays,
            sto_date(10),
            V_tabix TYPE sy-tabix,
            Y_tabix TYPE sy-tabix,
            sto_var1(10),
            sto_year(4),
            sto_month(2),
            sto_day(2),
            sto_final_date(10),
            W_HEADER LIKE LS_TARGET-RECORDTYPE,
            W_HEADER1(12) TYPE C VALUE 'HEDR00000000',
            W_FOOTER LIKE W_HEADER VALUE 'TRLR0000',
            CNT(5),
            CMD(125) TYPE C.
    **********CODE FOR GENRATING CSV FILE PATH*******************
    data: OUTFILE_NAME(100) TYPE C,
          OUTFILE_NAME1(10) TYPE C VALUE '/sapmnt/',
          OUTFILE_NAME3(18) TYPE C VALUE '/qoutsap/ATTUVS',
          DATE1 LIKE SY-DATUM,
          DD(2) TYPE C,
          MM(2) TYPE C,
          YYYY(4) TYPE C.
    MOVE SY-DATUM+6(2) TO DD.
    MOVE SY-DATUM+4(2) TO MM.
    MOVE SY-DATUM(4) TO YYYY.
    CONCATENATE YYYY MM DD INTO DATE1.
    CONCATENATE OUTFILE_NAME1 SY-SYSID OUTFILE_NAME3 '.CSV' INTO
    OUTFILE_NAME.
    **********END OF CODE FOR GENRATING CSV FILE PATH*************
      OPEN DATASET OUTFILE_NAME FOR OUTPUT IN TEXT MODE ENCODING DEFAULT.
    Code for generating Header.
      CONCATENATE W_HEADER1  SY-DATUM SY-UZEIT INTO W_HEADER.
      APPEND W_HEADER TO ET_TARGET.
      TRANSFER W_HEADER TO OUTFILE_NAME.
      CLEAR W_HEADER.
    End of code for generating Header.
    code for excluding the rows who's Quantity(PIECES) equal to zero.
      LOOP AT it_source INTO ls_source where KYF_0001 NE '0'.
    end of code for excluding the rows who's Quantity(PIECES) equal to
    *zero
        MOVE-CORRESPONDING ls_source TO ls_target.
        ls_target-RECORDTYPE = 'PKUP'.
        ls_target-CONTAINER = ''.
        ls_target-TRACKINGNUMBER = ls_source-ZINT_HU__Z_WM_HU.
        ls_target-PO = ''.
    Date Conversion for Staged Date.
        var1 = ls_source-CREATEDON.
        year = var1+0(4).
        month = var1+4(2).
        day = var1+6(2).
        CONCATENATE month '/' day '/' year INTO date.
    End of Date Conversion for Staged Date.
        ls_target-STAGEDDATE = date.
        ls_target-MOVEMENTTYPE = 'P'.
        ls_target-ORIGINCONTACTNAME = ''.
        ls_target-ORIGINCONTACTPHONE = ''.
        ls_target-DESTINATIONCONTACTNAME = ''.
        ls_target-DESTINATIONCONTACTPHONE = ''.
        ls_target-RCCCODE = ''.
        ls_target-GLCORCLLICODE = ''.
        ls_target-JFCCODE = ''.
        ls_target-DESCRIPTIONOFWORK1 = ''.
        ls_target-DESCRIPTIONOFWORK2 = ''.
        ls_target-INSTRUCTIONS = ''.
        ls_target-REQUESTEDSHIPDATE = date.
    Calculating STO Creation Date + 3 working Days.
        select /BIC/ZWORKDAYS from /bic/pzworkdays into table it_workdays.
        loop at it_workdays into wa_workdays.
            if  wa_workdays-/bic/zworkdays = ls_source-CREATEDON.
                V_tabix = sy-tabix.
                Y_tabix = V_tabix + 3.
            endif.
            If sy-tabix = y_tabix.
                sto_date = wa_workdays-/bic/zworkdays.
            endif.
        Endloop.
        clear v_tabix.
        clear Y_tabix.
        sto_var1 = sto_date.
        sto_year = sto_var1+0(4).
        sto_month = sto_var1+4(2).
        sto_day = sto_var1+6(2).
        CONCATENATE sto_month '/' sto_day '/' sto_year INTO sto_final_date.
    End Of Calculating STO Creation Date + 3 working Days.
        ls_target-REQUESTEDDELIVERYDATE = sto_final_date.
        ls_target-ATTSEORDER = ''.
        ls_target-CUBE = ''.
        ls_target-PIECES = ls_source-KYF_0001.
        ls_target-REEL = ''.
        ls_target-REELSIZE = ''.
        ls_target-VENDORSKU = ''.
        ls_target-ATTSESKU = ''.
        ls_target-COMPANYNAME = 'AT&T'.
        ls_target-OEM = ''.
        ls_target-REFERENCENUMBER2 = '0'.
        ls_target-REFERENCENUMBER3 = '0'.
        ls_target-REFERENCENUMBER4 = '0'.
        APPEND ls_target TO et_target.
        TRANSFER ls_target TO OUTFILE_NAME.
        CNT = CNT + 1.
      ENDLOOP.
        CNT = CNT + 2.
    Code for generating Header -Footer.
      SHIFT CNT LEFT DELETING LEADING SPACE.
      CONCATENATE W_FOOTER CNT INTO W_HEADER.
      APPEND W_HEADER TO ET_TARGET.
    End of code for generating Header -Footer.
    Code for file permissions
      TRANSFER W_HEADER TO OUTFILE_NAME.
      CLOSE DATASET OUTFILE_NAME.
      CONCATENATE 'chmod 644' OUTFILE_NAME INTO CMD SEPARATED BY SPACE.
      CALL 'SYSTEM' ID 'COMMAND' FIELD CMD.
    End of code for file permissions
    End of transformation code -
    ENDFORM.
    Hope it helps
    bhaskar

  • Open Hub export of file without meta file?

    Hi Gurus,
    I am daily exporting delta files with Open Hub Services to server.
    The name of the files include dates, so every day a new file is created and also a new meta S_ file is created.
    How can I export with Open Hub just data file without meta data file?
    Because every month there will be 30 meta files which will be of no use for me.....
    Regards, Uros

    Hi gurus,
    I've the same problem. I don't need of the S_ file. How can I block this file estraction??
    Thanks in advance & Best Regards,
    Domenico.

  • Access Violation when trying to read from Open Hub in BW

    Hi,
    Can someone help me out. I use BODS version 14.1.1.210. I have setup the connection to BW source and imported metadata for my OpenHub - no problem.
    Then created DTP and processchain to load Open hub data in BW.
    Made dataflow in BODS with Open Hub Tables as source and maintained processchain to be executed. When executing BODS job I get the following error:
    6800 10580 JOB 17-03-2014 13:10:34 Job <Test_OpenHub> is terminated due to error <170101>.
    6800 10580 SYS-170101 17-03-2014 13:10:34 |Session Test_OpenHub|Data flow Test Openhub
    6800 10580 SYS-170101 17-03-2014 13:10:34 System Exception <ACCESS_VIOLATION> occurred. Process dump option is off. Process is not dumped.
    6800 10580 SYS-170101 17-03-2014 13:10:34 Call stack:
    6800 10580 SYS-170101 17-03-2014 13:10:34 (null)
    6800 10580 SYS-170101 17-03-2014 13:10:34 Registers:
    6800 10580 SYS-170101 17-03-2014 13:10:34 RAX=00000000747E1648  RBX=0000000000209AB8  RCX=0000000000209AB8  RDX=0000000000000160  RSI=000000000231DAC0
    6800 10580 SYS-170101 17-03-2014 13:10:34 RDI=0000000008BEBAC0  RBP=0000000000000000  RSP=00000000002099C0  RIP=000000007477730A  FLG=0000000000010206
    6800 10580 SYS-170101 17-03-2014 13:10:34 R8=0000000000000000   R9=00000000FFFFFFFF  R10=0000000000620064  R11=0000000008C93600  R12=000000000033D390
    6800 10580 SYS-170101 17-03-2014 13:10:34 R13=0000000081D6FEC0  R14=000000000020A008  R15=000000000020A030
    6800 10580 SYS-170101 17-03-2014 13:10:34 Exception code: C0000005 ACCESS_VIOLATION
    6800 10580 SYS-170101 17-03-2014 13:10:34 Fault address:  000000007477730A 01:000000000002630A
    6800 10580 SYS-170101 17-03-2014 13:10:34 C:\Windows\WinSxS\amd64_microsoft.vc80.crt_1fc8b3b9a1e18e3b_8.0.50727.4940_none_88df89932faf0bf6\MSVCP80.dll
    6800 10580 SYS-170101 17-03-2014 13:10:34 ==========================================================
    6800 10580 SYS-170101 17-03-2014 13:10:34 Collect the following and send to Customer Support:
    6800 10580 SYS-170101 17-03-2014 13:10:34 1. Log files(error_*, monitor_*, trace_*) associated with this failed job.
    6800 10580 SYS-170101 17-03-2014 13:10:34 2. Exported ATL file of this failed job.
    6800 10580 SYS-170101 17-03-2014 13:10:34 3. DDL statements of tables referenced in this failed job.
    6800 10580 SYS-170101 17-03-2014 13:10:34 4. Data to populate the tables referenced in the failed job. If not possible, get the last few rows (or sample of them) when
    6800 10580 SYS-170101 17-03-2014 13:10:34 the job failed.
    6800 10580 SYS-170101 17-03-2014 13:10:34 5. Core dump, if any, generated from this failed job.
    6800 10580 SYS-170101 17-03-2014 13:10:34 ==========================================================
    Any good ideas?

    Hello Erik,
    not sure about it, but did you see this post some hours after yours?
    DS 4.1 job aborts with Access Violation
    Also, did you check ST22 to see if it generates any dump?
    And if you're able to check the BODS log files, it might me helpful. They're located at %LINK_DIR%\log directory. Check the names with the execution of the job.
    Hope it helps!
    Regards.
    Bruna Dupim

  • Open Hub does not work after upgrade

    We just upgraded our BW production system from 3.0b into 3.5 and we are running support package level 09. Now we are facing serious problems with Open Hub.
    After upgrade Open Hub delivering data from infocube into
    csv-files does not work. When running the InfoSpoke it
    will deliver 0 data records even though it should bring considerable amount of data.
    There is no error message or short dump, it just do not deliver any data. InfoCube has data in place and it is available for reportoing. Extraction is made in FULL-mode and there is no transformations. Funny thing is that extraction from ODS or InfoObject works perfectly fine.
    Has anyone idea how to solve this?
    Thanks in advance,
    JL

    We ran into the same issue as Jari Laine today.
    We just upgraded our BW sandbox environment from 3.0b into 3.5. We are running support package level 13, though. Now we are facing serious problems with Open Hub.
    After the upgrade, our Open Hub which delivers data from an info cube into CSV-files does not work any more.
    When running the Info Spoke it will deliver 0 data records even though it should bring a considerable amount of data.
    There is no error message or short dump, it just does not deliver any data. The Info Cube has data in place and it is available for reporting.
    The Extraction is made in FULL-mode and there are no transformations.
    Has anyone an idea how to solve this?
    Thanks in advance,
    Marc

  • Open Hub services in BI 7.0

    Hi Experts,
    I am working on BI 7.0 and have a requirement where i have to extract data from a standard sales cube(0SD_C03).I will be using open hub services,but i also need to send the master data attributes along with this into the application file server.
    I did the following steps
    1) created an openhub destinationa as a CSV file and selected the fields i needed.(Cube has 80+,but i selected only 10 fields for my file).
    2) added an infoobject YINVGP which is an attribute of 0Material
    3) The transformations were automatically created by the system.
    4) Created a tranformation between 0MATERIAL and its Attribute YINVGP and selected the rule type as "Read Master Data".
    5) When i click on transfer values to activate the changes,its giving me the following error.
    Rule11(Target/BIC/YINVGP Group:01 Standard group) No master data characterstic selected.
    Rule(Target Field:/BIC/YINVGP) No source assigned.
    Can the experts confirmif i am missing anything?If so please suggest on how to get the master data using open hub services.
    Is this possible with open hub services,if so i would really appreciate anybody can pass me screenshots or step by step procedure,as i have not worked on open hub earlier.
    thanks in advance

    Hi Naga,
    I have seen the link you have sent me.It doesnt really help me.But thanks anyways.I am able to create a CSV file,which is not a problem.But my question is can i get the attributes of 0Material to appear in the same file.SAP has provided an option to read master data in the transformation rules,and when i try to use the rule,its giving me an error saying source has not been assigned.
    thanks

Maybe you are looking for