UPLOAD A DIRECTORY WITH N NUMBER OF DATA FILE

Hi Experts
                           In my scenario i need to upload a directory with n number of excel file from my presentation server to application server . Please do help on that (note : directory with many file not an single file )

Hi,
Here is the Code for List of Files from the Specific Directory to Application Server(SAP)..
REPORT  ZDIRFILES    .
PARAMETER: p_fdir            type pfeflnamel DEFAULT '/usr/sap/tmp'.
data: begin of it_filedir occurs 10.
        include structure salfldir.
data: end of it_filedir.
*START-OF-SELECTION
START-OF-SELECTION.
Get Current Directory Listing for OUT Dir
  call function 'RZL_READ_DIR_LOCAL'
       exporting
            name     = p_fdir
       tables
            file_tbl = it_filedir.
List of files are contained within table it_filedir
  loop at it_filedir.
    write: / it_filedir-NAME.
  endloop.
Girish

Similar Messages

  • How to increase number of data files parameter

    Dear All,
    I am facing problem in increasing number of data files associated with the DB. I have DB_FILES = 200 and want to set it to 300. DB was started with SPFILE.
    Tell me how to increase this parameter. I dont want to stop DB also. I will be thankful if anyone guide me in both ways to stop DB and without stopping it.
    Regards
    Aqeel Nawaz

    Madrid,
    Please don't take it as nit-picking.Just wanted to point an enhancement in 10g regarding recreating control file for increasing number of datafiles.I am sure you must be aware about it.
    Please see this link and this note,
    http://download.oracle.com/docs/cd/B19306_01/server.102/b14231/dfiles.htm#sthref1340
    When you issue CREATE DATABASE or CREATE CONTROLFILE statements, the MAXDATAFILES parameter specifies an initial size of the datafile portion of the control file. However, if you attempt to add a new file whose number is greater than MAXDATAFILES, but less than or equal to DB_FILES, the control file will expand automatically so that the datafiles section can accommodate more files.
    From 10g onwards,the requirement to recreate teh control file is lifted.
    Please correct me if I interpreted the note in a wrong way.
    Regards
    Aman....

  • Change number of data file

    I need to copy the data files to a SAN and want to change the number of data files form 4 to 1. This is only a small database (22Gb). I don't know how to delete the data files and re create them.
    Regards
    Hans

    Hi Hans,
    no, you can also do it using DBMGUI. The process is really the same as when you're doing a homogenous system copy, the only difference is the target itself is the same db instance.
    You'd have to make sure you've backup up all (data+log) and no further data was created (for example by making a full data backup in ADMIN mode and saving the log). Then you'd need to initialize the DB and specify the desired number of (data) volumes and make sure you select the radio button labelled 'Create instance for recovery' during the initialization process, so the instance will only be put into ADMIN, after which you can then recover (using the backups made in the beginning).
    I'd recommend to test this out yourself on a small (non production) db instance.
    In comparison, from MaxDB version 7.6.00 Build 22 onwards, you would just have deleted the unwanted volumes (taking into respect that the left over volume can hold all data).
    Regards,
    Roland

  • The maximum number of data files that can be accessed on the ORACLE DB is r

    Hello Gurus,
    We are facing problem in our BW production server
    The maximum number of data files that can be accessed on the ORACLE DB is reached
    Error Message :
    ORA-00059: maximum number of DB_FILES exceeded
    We have search a SAP Note regarding this error but this is for oracle - 8i, we want to solve this problem in Oracle - 10g
    Thanks & Regards
    Shishir
    Vital Information
    Database - ORACLE - 10.2.0.2.0
    O/S - AIX - 5.3
    SAP - BIW ( 700 )

    Hi,
    check Data transfer from one BW system to another
    regards,
    kaushal

  • Sales report with Batch number & expiry dates

    Hello friends,
    would any one please help me to frame this query to achieve the above ..
    my query is as bellow, but I want to add the batch number & expiry dates along with the other information.....
    SELECT T0.[DocNum], T0.[DocDate], T0.[CardCode], T0.[CardName], T0.[NumAtCard], T1.[ItemCode], T1.[Dscription], T1.[Quantity] FROM OINV T0  INNER JOIN INV1 T1 ON T0.DocEntry = T1.DocEntry WHERE T1.[ItemCode] >=[%0] AND  T1.[ItemCode] <=[%1] AND  T0.[DocDate] >=[%2] AND  T0.[DocDate] <=[%3]
    So which table to join to get the batch & expiry...
    Rgds
    Suman

    HI
    Batch table is OIBT, IBT1,OBTN
    SELECT distinct T0.[DocNum], T0.[DocDate], T0.[CardCode], T0.[CardName], T0.[NumAtCard], T1.[ItemCode], T1.[Dscription], T1.[Quantity], T4.[ExpDate] FROM OINV T0  INNER JOIN INV1 T1 ON T0.DocEntry = T1.DocEntry inner join
    IBT1 T2 on T1.ItemCode = t2.ItemCode and T0.objtype=T2.basetype and T2.baseentry=T1.Docentry inner join
    oitm T3 on T1.Itemcode=T3.Itemcode inner join OIBT T4 on T1.Itemcode=T4.Itemcode  WHERE T1.[ItemCode] >=[%0] AND  T1.[ItemCode] <=[%1] AND  T0.[DocDate] >=[%2] AND  T0.[DocDate] <=[%3]
    group by T0.[DocNum], T0.[DocDate], T0.[CardCode], T0.[CardName], T0.[NumAtCard], T1.[ItemCode], T1.[Dscription], T1.[Quantity],T4.[ExpDate]
    Thanks
    Mohammad Imran

  • How to create line hart with variable number of data series

    Hello,
    I am student and I am completely new to Flex programming but
    I need to urgently create an app that should have a variable number
    of data series but of ame type of object.
    eg. Profits for X,Y and Z in in run while just corp X in the
    next instance.
    How can I accomodate for this in the app? can someone point
    me in the right direction I would really greatful as this project
    is really crucial for my coursework to get finished in timely
    manner.
    Thank you

    Hey,
    I'm not really sure what you want, but you should check out:
    http://demo.quietlyscheming.com/ChartSampler/app.html
    Which has a list of different types of charts in Flex and
    will hopefully help. Also check out charting from ILOG, which is
    quite good:
    http://www.ilog.com/products/elixir/

  • OraRRP Error with "Unable to copy data file;Error code 2, check disk space"

    Hi,
    Some users get this message -"Unable to copy data file;Error code 2, check disk space" when run report with orarrp, but most users do not get it.
    I check free space at both server and client side, they are very sufficient.
    I also checked directory exists for REPORTXX_TMP variable.
    My user call reports via URL (rwservlet) and it occur for all reports.
    How I can solve this problem?
    Thanks in advance.
    Tawatchai R.

    Hi,
    have the same problem now. One user has temporarily problems to download .rrpa files via URL (rwservlet) request. Error code: -"Unable to copy data file;Error code 2, check disk space". Did you get a solution??
    Thanks in advance. Axel

  • How can I read a binary file stream with many data type, as with AcqKnowledge physio binary data file?

    I would like to read in and write physiological data files which were saved by BioPac�s AcqKnowledge 3.8.1 software, in conjunction with their MP150 acquisition system. To start with, I�d like to write a converter from different physiodata file format into the AcqKnowledge binary file format for version 3.5 � 3.7 (including 3.7.3). It will allow us to read different file format into an analysis package which can only read in file written by AcqKnowledge version 3.5 � 3.7 (including 3.7.3).
    I attempted to write a reader following the Application Note AS156 entitled �AcqKnowledge File Format for PC with Windows� (see http://biopac.com/AppNotes/ app156Fi
    leFormat/FileFormat.htm ). Note the link for the Mac File format is very instructive too - it is presented in a different style and might make sense to some people with C library like look (http://biopac.com/AppNotes/ app155macffmt/macff.htm).
    I guess the problem I had was that I could not manage to read all the different byte data stream with File.vi. This is easy in C but I did not get very far in LabView 7.0. Also, I was a little unsure which LabView data types correspond to int, char , short, long, double, byte, RGB and Rect. And, since it is for PC I am also assuming the data to be written as �little endian� integer, and thus I also used byte swap vi.
    Two samples *.acq binary files are attach to this post to the list. Demo.acq is for version 3.7-3.7.2, while SCR_EKGtest1b.acq was recorded and saved with AcqKnowledge 3.8.1, which version number is 41.
    I would be grateful if you someone could explain how to handle such binary file stream with LabView and send an example to i
    llustrate it.
    Many thanks in advance for your help.
    Donat-Pierre
    Attachments:
    Demo.acq ‏248 KB
    SCR_EKG_test1b.acq ‏97 KB

    The reading of double is also straight forward : just use a dble float wired to the type cast node, after inverting the string (indian conversion).
    See the attached example.
    The measure of skin thickness is based on OCT (optical coherent tomography = interferometry) : an optical fiber system send and received light emitted to/back from the skin at a few centimeter distance. A profile of skin structure is then computed from the optical signal.
    CC
    Chilly Charly    (aka CC)
             E-List Master - Kudos glutton - Press the yellow button on the left...        
    Attachments:
    Read_AK_time_info.vi.zip ‏9 KB

  • Create table space command with no specification of data file path.

    I am using following command for creating table space in oracle 11g
    CREATE TABLESPACE testTbSpace DATAFILE 'dataFileName.dbf' SIZE 50M REUSE AUTOEXTEND ON NEXT 1M MAXSIZE 32767M NOLOGGING"
    But it is creating datafile dataFileName.dbf at disk at following path
    echo $ORACLE_HOME/dbs
    I dont want to create datafile at this path and also dont want to specify data file path in 'create table space' command.
    Is there is any parameter,which i can set and above command start to create dataFileName.dbf at that path
    Edited by: user8680179 on May 15, 2012 1:54 AM

    user8680179 wrote:
    i issued following commands from 'SYS' user;
    1. show parameter db_create_file_dest;
    NAME TYPE VALUE
    db_create_file_dest string
    2.alter system set db_create_file_dest='dataFilePath';
    System altered.
    3.show parameter db_create_file_dest;
    NAME TYPE VALUE
    db_create_file_dest string dataFilePath
    4.CREATE TABLESPACE testTbSpace2 DATAFILE 'test1.dbf' SIZE 50M REUSE AUTOEXTEND ON NEXT 1M MAXSIZE 32767M NOLOGGING;
    Tablespace created.
    But still my test1.dbf file is creating at old path($ORACLE_HOME/dbs)Is datafilepath a real location? I don't think so! Give a proper path like "d:\oracle\" and retry.
    Aman....

  • Number of data files per database!

    Hello,
    http://download.oracle.com/docs/cd/B19306_01/server.102/b14237/limits002.htm#i287915
    reads
    >>
    Database files
    Maximum per tablespace Operating system dependent; usually 1022
    Maximum per database 65533
    May be less on some operating systems
    Limited also by size of database blocks and by the DB_FILES initialization parameter for a particular instance
    >>
    I understand the DB_FILES part but how come number of files per database be limited by the database block size?
    The maximum number of datafiles per database is 65533 as per above quote.
    Could someone please explain this?
    Regards,

    Niten
    See the below link
    URGENT! - Data Packet Size
    Assign points if has really helped you.
    Regs
    Gopi

  • Problems with the number of image files

    Hi
    I am currently working in Quark and I have a big problem. Quark can't handle my 250 page, 700 image files document. If I try just scrolling through the documents from start to finish Quark crashes, and it doeesn't even create a rescue file. Now in the Quark forums they tell me to split the document into chapters. Chapters are fine if you are working with others on a projects as it lets different people work on different parts of the book at the same time. But this is all me, and I do A LOT of scrolling back and forth to cross check facts and so on. Having to constantly switch between files would be a huge pain. How would ID handle a document of this complexity?
    Sincerely
    Geir Rosset
    Oslo
    Norway

    But this is all me, and I do A LOT of scrolling back and forth to cross check facts and so on. Having to constantly switch between files would be a huge pain.
    I guess it's a matter of work styles, but I'd much prefer to switch around among a batch of open chapters with a simple key combination, as opposed to scrolling back and forth through pages in linear fashion.
    How would ID handle a document of this complexity?
    With the limited information you've offered, I'd say I have experience with similar documents, (product catalogs/pricelists), that ID handled just fine, and other cases in which seemingly simple files choked it. I can say the same of Quark. No one can answer your question definitively; there are too many variables. Chances are it's not (only) the size and complexity of your document in itself that are causing your problems. There can be any number of hardware or OS issues contributing, to name a couple.

  • Two instances/one running at a time with sae set of data files

    hi,
    I dont want to use OPS or Clustering...
    I want two diffrent instances ... with only one running at a time....
    When one will be down second will be manually started//
    Only requirement is that the set of datafiles must be same.........
    Is this possible...
    waiting for some inputs..
    thnx in advance

    Oracle FailSafe is just a set of services that hook into Windows clustering to allow you to have an active/passive cluster (i.e. one instance mounted at a time). It just saves you the effort of manually configuring a bunch of things that Windows provides out of the box. I don't believe there are licensing implications to using FailSafe, but your Oracle sales rep would be the definitive answer there.
    I'd strongly suggest that you use whatever clustering facilities are part of your OS. If you are trying to develop a general solution that will work on Windows and Solaris, I wish you luck. I'd bet it would be much easier to maintain a Windows-specific solution and a Solaris-specific solution using Windows- and Solaris-specific capabilities.
    Justin
    Distributed Database Consulting, Inc.
    http://www.ddbcinc.com/askDDBC

  • Error in running Extract Definition Upload from Data File concurrent.

    Hi all,
    Am trying to upload the 834 Extract Layout from the data file by running the concurrent program, Extract Definition Upload from Data File.
    After running this concurrent program am getting the Extract Layout definition with Record layout and Data elements within it.
    But some record layouts has the changes at their repeating level in that.
    Please suggest me how do i get the same repeating levels for the record layouts when i move the 834 benefit extract layout definition
    from one instance to another instance.

    Hi,
    We have exactly the same error in IBolt. This error happens sometime and we don't find the reason.
    IBolt has been upgraded to Version 3.1 SP1 and a fix has been developed by the support. After installation, we've work hardly during 2 days for testing all the flows by the customer without having this error.
    Ibolt works as a service and when this error occurs, the service stopped and must be restarted.
    In our flows, we have put a delete of the observer.dll file just before the rest of the flow but it doesn't solve the problem.
    Deleting the directory can be temporarly a solution but the error come back another day.
    Each time the SBO client or DTW is started, the %temp%\SMS_OBJ_DLL directory is created...
    When you just delete the observer.dll file and leave the rest of the directory like it is, the file is re-created when you start and log to SBO. I would say that this file is a copy of another file observer_800178.dll (800178 depending of the version of SBO yoi have). 800178 corresponds to SBO 2007 A PL 42.
    Do you have more information since your post has been send ?
    Thanks in advance for your help.
    Best regards.

  • Mass upload in KM maintaining the up-to-date versions of documents

    Hello guys,
    I need to do the Mass Upload in Repository Manager (persistence mode DB) in KM.
    My documents are stored in a file system, where they have versioning control. I need to upload the documents in portal with the up-to-date version.
    I have activated the versioning in the folder that I intend to upload the documents, but when I have done the upload, the documents have stayed with versions number = 1.
    Could you tell me what I need to do for upload the documents with the up-to-date version number, for example, if a document in file system has the version number 5, when I do the upload in the portal, the document should stay with the same version, 5 instead 1.
    Regards,
    Mary

    Hello Julian,
    Thanks for your help.
    There is only one document. There arenu2019t any old versions of document, the version of the document is kept by the author in the document, but only as a reference, to indicate that the document already was modified x times.
    The situation is like that:
    You open the Microsoft Word application and you created a document.
    You have the necessity to modify the document.
    You open the document and modify it and save it.
    Now the document has 2 versions, but the document is the same, no document is created from this.
    Inside the document in the second page, the author kept the versions, for example, the author creates a table where he specify the document version, like that:
    Version     Date            Author     
    01      13-08-2008        Mary
    02             14-08-2008        Mary
    I want to know what I need to do for when I upload the document in the portal, itu2019s appears with versions = 2 instead 1.
    Any idea about it.
    Regards
    Mary

  • Tungsten T3 - need to access DAT file with addresses - Please Help

    I have an old Palm Tungsten T3. The desktop computer with the synced address data has died and the T3 has lost power (so no addresses anymore).
    I have a .DAT file backup with all my addresses. Can anyone advise how I can get access to these.
    Reinstalling Palm Desktop 4.1.4e did not help as it doesn't allow me to import .DAT files.
    Can I import this file into any other app or open up in Excel etc?
    Any advice would be much appreciated.
    Thanks in advance.
    Post relates to: Tungsten T3
    This question was solved.
    View Solution.

    You have to physically copy the .dat file into the User directory over the existing address.dat file.  Find the one with your hotsync name abbreviated (i.e. - C:/Program Files/Palm (or PalmOne)/ your Hotsync name /address) and copy the .dat file there.
    Be sure you copy the file and keep another one safe in case something goes wrong.  The existing blank address.dat file will be only about 2k in size.
    Good luck!
    WyreNut
    I am a Volunteer here, not employed by HP.
    You too can become an HP Expert! Details HERE!
    If my post has helped you, click the Kudos Thumbs up!
    If it solved your issue, Click the "Accept as Solution" button so others can benefit from the question you asked!

Maybe you are looking for