Open Hub Extract in .txt file

Hi Expert,
A abc.txt file is getting populated as a target of open hub. That file is populated through a process chain. DTP monitor screen shows that output data volume is around 2.8 million everyday, but I am not been able to find data more than 1250000. Every time whatever be the data target in dtp monitor it shows me 1250000 in the txt file. Is there any limitation on the number of rows in .txt file.
Please help.
Regards,
Snehasish

thanks for the reply Geetanjali.
Yeah one field routine is there , but from dtp monitor screen i can see that total record count is 2608000 but when i m chking from the .txt file . it is showing me 1250000 irrespective of the dtp monitor record count. Moreover, I m just populating 3 column in my target txt file.

Similar Messages

  • Open Hub: How-to doc "How to Extract data with Open Hub to a Logical File"

    Hi all,
    We are using open hub to download transaction files from infocubes to application server, and would like to have filename which is dynamic based period and year, i.e. period and year of the transaction data to be downloaded. 
    I understand we could use logical file for this purpose.  However we are not sure how to have the period and year to be dynamically derived in filename.
    I have read in sdn a number of posted messages on a similar topic and many have suggested a 'How-to' paper titled "How to Extract data with Open Hub to a Logical Filename".  However i could not seem to be able to get document from the link given. 
    Just wonder if anyone has the correct or latest link to the document, or would appreciate if you could share the document with all in sdn if you have a copy.
    Many thanks and best regards,
    Victoria

    Hi,
    After creating open hub press F1 in Application server file name text box from the help window there u Click on Maintain 'Client independent file names and file paths'  then u will be taken to the Implementation guide screen > click on Cross client maintanance of file name > create a logical file path by clicking on new entiries > after creating logical file path now go to Logical file name definition there give your Logical file , name , physical file (ur file name followed by month or year what ever is applicable (press f1 for more info)) , data format (ASC) , application area (BW) and logical path (choose from F4 selection which u have created first), now goto Assignment of  physical path to logical path > give syntax group >physical path is the path u gave at logical file name definition.
    however we have created a logical path file name to identify the file by sys date but ur requirement seems to be of dynamic date of tranaction data...may u can achieve this by creating a variable. U can see the help from F1 that would be of much help to u. All the above steps i have explained will help u create a dynamic logical file.
    hope this helps u to some extent.
    Regards

  • Open hub error when generating file in application server

    Hi, everyone.
    I'm trying to execute an open hub destination that save the result as a file in the application server.
    The issue is: in production environment we have two application servers, XYZ is the database server, and A01 is the application server. When I direct the open hub to save file in A01 all is working fine. But when I change to save to XYZ I´m getting the following error:
    >>> Exception in Substep Start Update...
    Message detail: Could not open file "path and file" on application server
    Message no. RSBO214
    When I use transaction AL11, I can see the file there in XYZ filesystem (with data and time correspondent to execution), but I can´t view the content and size looks like be zero.
    Possible causes I already checked: authorization, disk space, SM21 logs.
    We are in SAP BW 7.31 support package 6.
    Any idea what could be the issue or where to look better?
    Thanks and regards.
    Henrique Teodoro

    Hi, there.
    Posting the resolution for this issue.
    SAP support give directions that solved the problem. No matter in which server (XYZ or A01) I logon or start a process chains, the DTP job always runs in A01 server, and it causes an error since the directory doesn´t exist in server XYZ.
    This occurs because DTP settings for background job was left blank. I follows these steps to solve the problem:
    - open DTP
    - go to "Settings for Batch Manager"
    - in "Server/Host/Group on Which Additional Processes Should Run" I picked the desired server
    - save
    After that, no matter from where I start the open hub extraction, it always runs in specified server and saves the file accordingly.
    Regards.
    Henrique Teodoro

  • To create an header for Open Hub extract

    Hi guys,
    How do i get an header for an open hub extract for a CSV file, an header consisting of the field names technical or otherwise.
    Thanks,
    Your help will be greatly appreciated

    hi,
    Field definition
    On the Field Definition tab page you define the properties of the fields that you want to transfer.
    We recommend that you use a template as a basis when you create the open hub destination. The template should be the object from which you want to update the data. This ensures that all the fields of the template are available as fields for the open hub destination. You can edit the field list by removing or adding fields. You can also change the properties of these fields.
    You have the following options for adding new fields:
    ●     You enter field names and field properties, independent of a template.
    ●     You select an InfoObject from the Template InfoObject column. The properties of the InfoObject are transferred into the rows.
    ●     You choose  Select Template Fields. A list of fields are available as fields for the open hub destination that are not contained in the current field list. You transfer a field to the field list by double-clicking on it. This allows you to transfer fields that had been deleted back into the field list.
    If you want to define the properties of a field so that they are different from the properties of the template InfoObject, delete the template InfoObject entries for the corresponding field and change the properties of the field. If there is a reference to a template InfoObject, the field properties are always transferred from this InfoObject.
    The file or database table that is generated from the open hub destination is made up of the fields and their properties and not the template InfoObjects of the fields.
    If the template for the open hub destination is a DataSource, field SOURSYSTEM is automatically added to the field list with reference to InfoObject 0SOURSYSTEM. This field is required if data from heterogeneous source systems is being written to the same database table. The data transfer process inserts the source system ID that is relevant for the connected DataSource. You can delete this field if it is not needed.
    If you have selected Database Table as the destination and Semantic Key as the property, the field list gets an additional column in which you can define the key fields for the semantic key.
    In the Format column, you can specify whether you want to transfer the data in the internal or external format. For example, if you choose External Format here, leading zeros will be removed from a field that has an ALPHA conversion routine when the data is written to the file or database table.
    Assign points if helpful
    Cheers!!
    Aparna

  • How can i open a DOC or TXT file and insert the data into table?

    How can i open a DOC or TXT file and insert the data into table?
    I have a doc file . the doc include some columns and some rows.(for example 'ID,Name,Date,...').
    I'd like open DOC file and I'd like insert them into the table with same columns.
    Thanks.

    Use the SQL*Loader utility or the UTL_FILE package.

  • Using open hub to create a file on another server

    Hello everyone,
    I am wanting open hub to create a file, but create it on a different server (a file sharing server not associated with SAP).  Both BW and this file share server run Windows.
    Can this be done using logical files  (FILE transaction) or other configuration?
    I am aware of some other people posting here that send the file from open hub to DIR_HOME and then have created an OS command (SM49) to copy to the other file share server.  Also I know one could also write an ABAP to FTP the file from this location out to the file share server as well.
    I was hoping that FILE transaction would allow this but so far I have not had any luck with it.
    Many Thanks,
    Kevin

    All,
    Found out that this can be done, you can send a flat file from BW to a completely different server in one step, no need for SM49 or AL11 or FTP or ABAP program in the middle.  The trick lies in the configuration in transaction FILE.
    FILE transaction configuration steps:
    Step 1 - Define Logical File Path - just a place holder here
    Step 2 - Assignment of Physical Paths to Logical Path, here is where you enter the server & path but not a filename, for example
    \\server_swvs01\SAPFILE\<SYSID>\BPC\Cust_R3_sales\<FILENAME>
    Step 3 - Define Cross-client Logical File Name, here is where you specify the actual file name to be written to the file share on the other server in field "Physical file", just make sure and link this to the Local path you created in Step 1 above.  You can also use system variables in the physical file name or hard code it all.
    your_file_name.csv
    Hope this helps someone
    Cheers,
    Kevin

  • Open Hub - Header in flat file

    Hi there Experts.
    I am working with a opeb Hub destination, and so far everything has been straight forward.
    But i need to include headers in the first line of the file.
    By default the Open Hub i generating a Control File, so it is possible to look in this file, but the requirements from the customer is to have fieldnames inside the same file as the data.
    Any help will be awarded.
    Ingrid

    I have faced this. There is no delivered mechanism. One work around would be (after the extraction) to open the file using ABAP and paste the header. If one tries to figure (in the start routine or end routine) that this is the first record of the first package, one faces the problem with data types (header has a CHAR data types where the fields of the file may not all be CHAR). We eventually convinced the user to drop the requirement. As you mention, there are generated files with header information.
    Not sure this helps.

  • Open hub destination problem - .csv file is empty

    Hi everyone,
    I am trying to extract data from our BW system using Open Hub Dest. to a .csv file
    DTP runs successfully means show total no. of records but when I see in AL11 then records are not in .csv file that is .csv file is empty.
    How to figure out the problem?
    Best regards
    Ahmad

    @Jagadeesh:
    if I correctly understood you, you mean to say that first I should run DTP and save to an excel file to see if it shows data or not. This i think is not possible; as when we create OHD then 'Destination Type' we can choose among 'DB table', 'table' and 'third party tool',
    so if I choose 'Table' then 'File Name' is .csv file (same name as the OHD) and can change it to .xls file
    @Arun:
    have permission to write in the folder
    Do you mean creating OH to a Database Table?
    in DTP I have a filter on a Characteristic, when its 1500 records then it works (data is in .csv file) but when 3500 records then .csv file is empty
    no error message, no log
    Regards
    Ahmad

  • Open Hub Destination Table Vs File

    Hello Experts,
    We have a requirement to send data from BW to 3rd party system every day. In the Open hub destination we have two options for the target.
    1. Table
    2. File
    ETL tool like Informatica will extract the data from table or file into 3rd party sytem.
    Which way is the best in performance and maintenance?
    Thanks in advance
    Sree

    Hi,
    If you follow the delta mechanism then table is good. You can also go for File not a problem. See the both eamples in the following articles.
    If you use Fiels then go for Application Server option, i.e. AL11 becasue no one can change teh files in this location, you can restrict the authorizations.
    Open Hub Destination: Part 1
    http://www.sdn.sap.com/irj/servlet/prt/portal/prtroot/com.sap.km.cm.docs/library/business-intelligence/m-o/open%20hub%20destination%3a%20part%201.pdf
    Open Hub Destination: Part 2
    http://www.sdn.sap.com/irj/servlet/prt/portal/prtroot/com.sap.km.cm.docs/library/business-intelligence/m-o/open%20hub%20destination%3a%20part%202.pdf
    Thanks
    Reddy

  • Problem with a Open Hub Extraction

    Hi Experts,
    I have a huge problem here.
    I´m trying  to extract some data with an Open Hub application, but i´m having trouble with the application server.
    See, I have to extract the information from a DSO to a file in a application server, but the company has one application server for each environment (BD1, BQ1, etc) and the conversion it´s not working and the transportation of the request always fails.
    Anyone knows how to fix this??
    Thanks.
    Regards.
    Eduardo C. M. Gonçalves

    Hi,
    While creating the open hub, you need to maintain the file name details under T-CODE - FILE.
    I hope you have maintained that.
    There also you wil have to use the variable to change the file path which will change the file path in each system..
    There you will have to define
    Logical File Path Definition
    Assignment of Physical Paths to Logical Path
    Logical File Name Definition, Cross-Client
    Once you have define this and transport , then your Open Hub will go smooth.
    Thanks
    Mayank

  • Open Hub export to ASCII file

    Is it possible to export infoprovider data through open hub to an ASCII formatted file? How would we do that in BI 7.0?

    Hi,
    You can use TCode RSANWB to goto APD and u can export data from any infoprovider to Flat File .
    Check this link on APD:
    http://help.sap.com/saphelp_nw04/helpdata/en/49/7e960481916448b20134d471d36a6b/frameset.htm
    http://help.sap.com/saphelp_nw04/helpdata/en/49/7e960481916448b20134d471d36a6b/frameset.htm
    https://www.sdn.sap.com/irj/servlet/prt/portal/prtroot/docs/library/uuid/96939c07-0901-0010-bf94-ac8b347dd541
    Rgds
    SVU123

  • Open Hub Extraction

    Hello
    I am doing a open hub and sending data to a file. Now in the file i want to maitain the header line(description of field) for the fields i am sending to the file.  How can i do that. Please can someone explain me a little bit in brief.
    thanks

    Hi,
    Once u run the Infospoke, it will generates two files, data file and schema file. The data file contains the data and the schema file contains the deader or the structure of the Infospoke. So if u want to maintain the header in the data file, u just copy the list of InfoObjects name and paste it on the header file.
    Thanks & Regards
    Ramakrishna Kamurthy

  • Open Hub extraction for Hierarchies

    Hello,
    I need to create a Open Hub on the 0MAST_CCTR in HR-PA. Unfortunately the Open Hub does not give you the option for hierarchies. Any ideas?
    Thanks,

    Hi,
      If it is a flat file output then you can use the below program mentioned in SAP how to document. We are using this in our system.
    http://www.sdn.sap.com/irj/scn/go/portal/prtroot/docs/library/uuid/0403a990-0201-0010-38b3-e1fc442848cb?quicklink=index&overridelayout=true
    Regards,
    Raghavendra.

  • Open Hub Services, Format of files

    Hi everybody
    Does anybody knows how to create a CSV file in ascii format using an open hub in a process chain ?
    On my system which is a unicode it seems to be as if I could only produce unicode file when I use the process chain.
    Thank you for your help
    Alex

    Hi,
    Load the data into cube as normal process of process chain, then the second step is u should create an info spoke using RSBO, dran and drop the info spoke after the cube data load, there while u r creating the Info spoke it will give u the option of file transfer whethere u want to tansfer it as a CSV , ASCII file.Hope this helps.

  • Open Hub Destination with Flat File - Structure file is not correct

    Using the new concept for Open Hub destination (NOT an InfoSpoke).
    Output is to a Flat File. Remember that when output is to a flat file 2 files are generated: the output file and a structure file (S_<name of output file>.CSV)
    The Output file structure corresponds to the structure defined in the Open Hub destination.
    However, the S_ (structure) file does not reflect the proper structure. The S_ file lists all fields sorted in alphabetical order based on the field name, instead of listing all fields in the correct order that they appear in the output file.
    I have not found any settings that I can modify for the S_ file. When using InfoSpokes the S_ file is correct and I do not have to do anything.
    The issue is that the S_ file cannot be used to describe the output file. This is a problem.
    Anyone knows a solution to this issue? The goal is to get the S_ file to reflect the proper structure.
    I could not find any Notes on this. Please do not send general help links, this is a specific question.
    Thank you all for your support.

    Hi,
    However, the S_ (structure) file does not reflect the proper structure. The S_ file lists all fields sorted in alphabetical order based on the field name, instead of listing all fields in the correct order that they appear in the output file.
    As I know and checked this is not the case. The S_ file has the fields in the order of infospoke object sequence ( and in transformation tab, source structure).
    I would suggest you to control it again.
    Derya

Maybe you are looking for

  • The server was unable to load the SSL provider library needed to log in

    Hello All, When I am trying to login to my default SQL 2008R2 SP2 instance in windows 2003 server via ssms, I am getting the below error A connection was successfully established with the server, but then an error occurred during the pre-login handsh

  • Time Machine v western digital my book live fail second backup

    HI I am using iMac 27 inch, running Lion. I used Western Digital Mybook live as my time machine backup. I follow the instrustion step by step, its work fine. However, when I restarted my mac. Then its start fail to do continues backup. I can't do any

  • Trying to fit output of XY Graph

    Hello, I'm trying to fit the output of an XY graph which is (1-D Array of Cluster of 2 Elements(2x 1-D Array of Double[64-bit Real])). It seems like instead of having one array with two columns X,Y indexed persample number, I want to have a 1-D Array

  • HT204406 Library stalls when uploading from Windows 7; starts over each time I open itunes.

    Running windows 7 pro; latest version of iTunes; when running "Match," it uploads all of my songs (after a few restarts) but then stalls on the "rest of your music" upload.  (Total of about 15,000 songs: 10,000 "iTunes" songs; about 5,000 of "your mu

  • MSI GX630 shuts down at bios

    When I start up my laptop, the BIOS screen shows up, then after a few seconds the laptop just shuts down every time. I've tried removing the hard drive and trying it, with the same result. I removed power and the battery for 30 seconds with the same