Unable to read large volume of file from PI to BW

Hi,
I am reading data from PI(XI) to BW using web service data source. The data reach the report if the xml contains less number of records (say 2000-3000) but it does not reach the report if the xml contains more number of data (say 10k).
I checked in sxmb_moni transaction and the process is always scheduled status and the flag is green.
Can somebody tell me asto what exactly the issue is?
The BW version is 7.0.
Thanks and Regards,
Arya

Hi,
I'm also trying to integrate bw using webservice DS, but the probem is that even several records can't reach further sxmb_moni (with green flag there). When I use FM (u201C/BIC/CQu201D + DSName) open request appears in RSRDA (yellow) and I can see it in PSA. Then open request in RDA Monitor for DTP should be created, but nothing happens. (Daemon started)
Could someone help me in such situation?
Thanx in advance,
Alex

Similar Messages

  • Transferring large volume of files from mac to PC?

    Hi i have a mac with osx 10.4.8 and a pc, i need to transfer a large amount of files from mac to pc (around 200GB)
    Now i have lots of external HDs which are all used by either the PC or the macs. the ones formatted for the macs cannot be read atall by the PC without some expensive software, the PC formatted one i have appears to be readable on the mac, but when i try to copy the files onto it or change anything, i am not able to change/add anythingn due to permissions problems.
    Not sure what to do, i have quite a few HDs with plenty of sapce, but if they are all in the wrong format i cant really wipe/reformat them easily with files on them, nor do i want to buy yet another HDD just to transfer some files....
    Any ideas/advice?

    https://discussions.apple.com/docs/DOC-3003

  • I am unable to read signatures on PDF files sent from my Los Angeles office - they use windows, any solution?

    I am unable to read signatures on PDF files sent from my Los Angeles office - they use windows, any solution?

    Hey guys,
    So this is follow up from my debarkle with the EDD. I found out my problem with copying files from Mac to EDD and vice versa was a result of a not so good EDD ( i had an apollo hard drive from imation) that was not very compatible with macs. So i did my research and found out that the best hard drives were Western Digital and Seagate. I bought the newest western digital EDD 1TB and formated it to FAT32 and guess what...no problems so far. The only problem is that FAT32 format doesn't copy files larger than about 4 gigs so i couldnt copy a movie from my brothers computer onto my EDD that was 1080p. You could probably resolve that by partitioning a small part of your hard drive in ExFAT? but yeah, hopefully that helped guys.
    Aaisha

  • BUG - Unable to open Canon CR2 RAW file from G1 X, ACR 7.0.0.308

    Unable to open Canon raw CR2 file from G1 X (Photoshop cannot open this file), CR2 file from 1DSIII opens ok.
    Windows 7 X64, Nvidia 2000D, Photoshop CS6 X64 BETA.

    That is correct. the G1 X will be supported in a 7.0 dot release. The camera is currently supported in Lightroom 4.
    Camera Raw plug-in | Supported cameras
    Note: These cameras are supported in Adobe Photoshop Lightroom 4.0, but not in Camera Raw 7.0.  These cameras will be supported in the next dot release of Camera Raw 7.

  • "Unable to read or process latte file" message when attempting to photo merge out of photoshop cc, bridge, essentials organizer or editor.  Search of "latte file" brings no explanation on the web.  What's going on?  Help?

    "Unable to read or process latte file" message when attempting to photo merge out of photoshop cc, bridge, essentials organizer or editor.  Search of "latte file" brings no explanation on the web.  What's going on?  Help?

    From the web:
    Templating | Nette Framework
    "One of the best template engines is part of Nette Framework and it is called Latte."
    Were you working on a web document or something?
    I have no clue myself.

  • How to copy very large number of files from one drive to another???

    I'm a fairly experienced Mac user for serveral years but this problem really has me stumped.
    I'm trying to copy or move 152,000 files from one external drive to another drive. I can highlight (Cmd - A) all the files on the first drive and drag them to the second drive but Finder always shows 32,768 files being copied no matter what I try.
    Any and all suggestions on how to move/copy a large number of files from one external drive to another are greatefully appreciated.
    Thank you in advance,
    Mack

    I would use the command line tool rsync.
    For instance with: rsync -av source-dir destination-dir
    -a The files are transferred in "archive" mode, which ensures that symbolic links, devices, attributes, permissions, ownerships, etc. are preserved in the transfer.
    -v Verbose, so you see the progress.
    Rsync is fast and really, really powerful and many times used in shell scripts and the like to automatically backup and/or sync stuff. Google a bit for more info.

  • How to read/write a binary file from/to a table with BLOB column

    I have create a table with a column of data type BLOB.
    I can read/write an IMAGE file from/to the column of the table using:
    READ_IMAGE_FILE
    WRITE_IMAGE_FILE
    How can I do the same for other binary files, e.g. aaaa.zip?

    There is a package procedure dbms_lob.readblobfromfile to read BLOB's from file.
    http://download-east.oracle.com/docs/cd/B19306_01/appdev.102/b14258/d_lob.htm#sthref3583
    To write a BLOB to file you can use a Java procedure (pre Oracle 9i R2) or utl_file.put_raw (there is no dbms_lob.writelobtofile).
    http://asktom.oracle.com/pls/ask/f?p=4950:8:1559124855641433424::NO::F4950_P8_DISPLAYID,F4950_P8_CRITERIA:6379798216275

  • How do I set up my Mac mini to read iTunes and iPhoto files from my new Mac mini server?

    Want my old Mac mini to read iPhoto and iTunes files from my new Mac mini server.  Old Mac has no more room to store files.  How do I do this?

    Hello,
    Have you tried enabling iTunes Sharing & iPhoto Sharing on the Server?
    Or maybe...
    http://support.apple.com/kb/ht1449
    http://www.apple.com/itunes/inside-itunes/2011/02/how-to-move-your-itunes-librar y-to-another-hard-drive.html
    http://support.apple.com/kb/PH2506
    http://support.apple.com/kb/ht1229

  • SSIS 2008 – Read roughly 50 CSV files from a folder, create SQL table from them dynamically, and dump data.

    Hello everyone,
    I’ve been assigned one requirement wherein I would like to read around 50 CSV files from a specified folder.
    In step 1 I would like to create schema for this files, meaning take the CSV file one by one and create SQL table for it, if it does not exist at destination.
    In step 2 I would like to append the data of these 50 CSV files into respective table.
    In step 3 I would like to purge data older than a given date.
    Please note, the data in these CSV files would be very bulky, I would like to know the best way to insert bulky data into SQL table.
    Also, in some of the CSV files, there will be 4 rows at the top of the file which have the header details/header rows.
    According to my knowledge I would be asked to implement this on SSIS 2008 but I’m not 100% sure for it.
    So, please feel free to provide multiple approaches if we can achieve these requirements elegantly in newer versions like SSIS 2012.
    Any help would be much appreciated.
    Thanks,
    Ankit
    Thanks, <b>Ankit Shah</b> <hr> Inkey Solutions, India. <hr> Microsoft Certified Business Management Solutions Professionals <hr> http://ankit.inkeysolutions.com

    Hello Harry and Aamir,
    Thank you for the responses.
    @Aamir, thank you for sharing the link, yes I'm going to use Script task to read header columns of CSV files, preparing one SSIS variable which will be having SQL script to create the required table with if exists condition inside script task itself.
    I will be having "Execute SQL task" following the script task. And this will create the actual table for a CSV.
    Both these components will be inside a for each loop container and execute all 50 CSV files one by one.
    Some points to be clarified,
    1. In the bunch of these 50 CSV files there will be some exception for which we first need to purge the tables and then insert the data. Meaning for 2 files out of 50, we need to first clean the tables and then perform data insert, while for the rest 48
    files, they should be appended on daily basis.
    Can you please advise what is the best way to achieve this requirement? Where should we configure such exceptional cases for the package?
    2. For some of the CSV files we would be having more than one file with the same name. Like out of 50 the 2nd file is divided into 10 different CSV files. so in total we're having 60 files wherein the 10 out of 60 have repeated file names. How can we manage
    this criteria within the same loop, do we need to do one more for each looping inside the parent one, what is the best way to achieve this requirement?
    3. There will be another package, which will be used to purge data for the SQL tables. Meaning unlike the above package, this package will not run on daily basis. At some point we would like these 50 tables to be purged with older than criteria, say remove
    data older than 1st Jan 2015. what is the best way to achieve this requirement?
    Please know, I'm very new in SSIS world and would like to develop these packages for client using best package development practices.
    Any help would be greatly appreciated.
    Thanks, <b>Ankit Shah</b> <hr> Inkey Solutions, India. <hr> Microsoft Certified Business Management Solutions Professionals <hr> http://ankit.inkeysolutions.com
    1. In the bunch of these 50 CSV files there will be some exception for which we first need to purge the tables and then insert the data. Meaning for 2 files out of 50, we need to first clean the tables and then perform
    data insert, while for the rest 48 files, they should be appended on daily basis.
    Can you please advise what is the best way to achieve this requirement? Where should we configure such exceptional cases for the package?
    How can you identify these files? Is it based on file name or are there some info in the file which indicates
    that it required a purge? If yes you can pick this information during file name or file data parsing step and set a boolean variable. Then in control flow have a conditional precedence constraint which will check the boolean variable and if set it will execute
    a execte sql task to do the purge (you can use TRUNCATE TABLE or DELETE FROM TableName statements)
    2. For some of the CSV files we would be having more than one file with the same name. Like out of 50 the 2nd file is divided into 10 different CSV files. so in total we're having 60 files wherein the 10 out of 60 have
    repeated file names. How can we manage this criteria within the same loop, do we need to do one more for each looping inside the parent one, what is the best way to achieve this requirement?
    The best way to achieve this is to append a sequential value to filename (may be timestamp) and then process
    them in sequence. This can be done prior to main loop so that you can use same loop to process these duplicate filenames also. The best thing would be to use file creation date attribute value so that it gets processed in the right sequence. You can use a
    script task to get this for each file as below
    http://microsoft-ssis.blogspot.com/2011/03/get-file-properties-with-ssis.html
    3. There will be another package, which will be used to purge data for the SQL tables. Meaning unlike the above package, this package will not run on daily basis. At some point we would like these 50 tables to be purged
    with older than criteria, say remove data older than 1st Jan 2015. what is the best way to achieve this requirement?
    You can use a SQL script for this. Just call a sql procedure
    with a single parameter called @Date and then write logic like below
    CREATE PROC PurgeTableData
    @CutOffDate datetime
    AS
    DELETE FROM Table1 WHERE DateField < @CutOffDate;
    DELETE FROM Table2 WHERE DateField < @CutOffDate;
    DELETE FROM Table3 WHERE DateField < @CutOffDate;
    GO
    @CutOffDate which denote date from which older data have to be purged
    You can then schedule this SP in a sql agent job to get executed based on your required frequency
    Please Mark This As Answer if it solved your issue
    Please Vote This As Helpful if it helps to solve your issue
    Visakh
    My Wiki User Page
    My MSDN Page
    My Personal Blog
    My Facebook Page

  • Find and Read the Target Spreadsheet File from Target Folder

    I'm trying to find and read the target spreadsheet file from the target folder. Please help. Thanks.
    Solved!
    Go to Solution.

    If you know the folder, you could do a List Folder Contents to get a list of files in that folder.  You can then just use Array Index to get the file you want.
    There are only two ways to tell somebody thanks: Kudos and Marked Solutions
    Unofficial Forum Rules and Guidelines

  • Read and List the Files from Remote Webserver Path

    Hi All,
    I have requirement where i need to Read and List the files from a Folder of Remote webserver path using JAVA.
    Remote webserver is within the network only...No Firewall and also Access is given to Read the folder. No Issues on this.
    Folder will just contain some PDF files...
    I just need to display the PDF file names available in the Folder..
    No need to read the PDF File...Only required to read the folder to list the file names in it.
    Looking forward some workaround to this.
    Thanks and Regards.

    I need to read the folder from a webserver path of different machine...
    File dir = new File( prop.getProperty("inputPath"));
    File[] files = dir.listFiles(fileFilter);     
    final String match=siteName;
         final String type=reportType;
         Calendar c1 = Calendar.getInstance();
         c1.add(Calendar.MONTH, - Integer.parseInt(prop.getProperty("filterMonths"))); //Filters reportes generated in last X months (X picked from config file)           
         final long filterDate = c1.getTime().getTime();           
         FileFilter fileFilter = new FileFilter() {
         public boolean accept(File file) {
         long fileLastModiDate = file.lastModified();      
         if((fileLastModiDate >= filterDate) && (file.getName().toLowerCase().startsWith(match)) && ( (type.equals("M") && file.getName().indexOf("WIP")==-1) || ((!type.equals("M") && file.getName().indexOf("WIP")!=-1)) ) ) {                        
              return true;
         }else {
              return false;
    Here it works fine if the input path is local machine..
    But i need to know how to give the input path as WEBSERVER PATH of different machine??

  • 043: Unable to read the Software Rules file

    hi,
    I have the message "043: Unable to read the Software Rules file " that
    appears on workstation object under Inventory tab, my software rule is
    activated, their is no other options to activate for this rule, is there
    other Nds objects that I need to configure for software scans?
    thanks!

    Dominic,
    It appears that in the past few days you have not received a response to your posting. That concerns us, and has triggered this automated reply.
    Has your problem been resolved? If not, you might try one of the following options:
    - Do a search of our knowledgebase at http://support.novell.com/search/kb_index.jsp
    - Check all of the other support tools and options available at http://support.novell.com in both the "free product support" and "paid product support" drop down boxes.
    - You could also try posting your message again. Make sure it is posted in the correct newsgroup. (http://support.novell.com/forums)
    If this is a reply to a duplicate posting, please ignore and accept our apologies and rest assured we will issue a stern reprimand to our posting bot.
    Good luck!
    Your Novell Product Support Forums Team
    http://support.novell.com/forums/

  • I am trying to resize disk space on parallels 6 for windows XP under configure Hardrive and I get the error resizing failed unable to resize last volume the file system of this volume might be damaged or not supported?  I am on a Mac using Parallels 6

    I have been trying to increase my disk space for windows xp on my Imac and I am running parallels 6.  before opening windows I go into configure in parallels and open the hard drive and try to move the allotted disk space from 32 gig to 64 or any other increased gig and then apply and I get the error message: resizing failed unable to resize last volume.  The file system of this volume might be damaged or not supported.  How do I increase the size??  I just also did a disk repair utility with Apple on the Apple side of my IMAC.  Please Help!!   I had uninstalled SQL server on the windows side and needed to uncompress files to install this and then I ran short on disk space.

    This sounds to me that your Parallels XP file system is corrupted and not the OSX file system.
    Have you checked the XP file system while running the Parallels XP yet ?
    However, it would be better to ask such question in the Parallels Forum http://forum.parallels.com/
    Stefan

  • Permissions "read only" after copying files FROM server

    Hi everyone!
    OK, just setup a Leopard server. Clients all at Tiger - level..for now. This is what I want to do:
    I want a read only - network library drive so that the clients can grab files from but then modify locally on their machines. No need to put back on that server as I have another system in place for things like that.
    *The problem:* If I propagate permissions at "read only" then when the user copies to their local drives, the owner changes but the permissions remain. They have to manually change the permissions back to match their local, then they can work with the file. When we're talking about hundreds of files for 15+ users a day, then it is quite a pain and a show stopper.
    I had this setup previously on a panther server with no trouble. I've maintained settings as close as possible but still having trouble.
    Any ideas as to what I'm missing?
    FYI- I don't want to change permissions to read/write to protect the library from tampering.
    thanks

    +If I propagate permissions at "read only" then when the user copies to their local drives, the owner changes but the permissions remain.+
    OK, from your description, you'd have to have the POSIX owner's permissions set to just read-only or read and execute. As you've observed, a copy obtains permissions in the following manner:
    The POSIX owner changes to match the account performing the copy (the action-bearing account), and the POSIX group is inherited from the copy's destination parent. The POSIX permissions stay the same, though. (Inherited ACL entries on the original are also lost; explicitly-defined ones are preserved, and new inherited entries apply to the copy from its destination parent.)
    All of this is proper behavior, and there's an easy way to make it work to suit your needs.
    Since the POSIX owner changes to the action-bearing account for the copy, you'll want to ensure that the original POSIX owner's permissions are read and write. This does not require that you make the original's POSIX owner the same as that of the action-bearing account. An example would better illustrate the situation:
    Example POSIX Permissions for the Original on the server:
    Owner root can read and write
    Group somegroup can only read
    Everyone else can only read
    Thus, let's say that tsmith is the action-bearing account. So tsmith copies the file to his desktop, and the POSIX permissions now look like this:
    Owner tsmith can read and write
    Group staff can only read
    Everyone else can only read
    Thus, tsmith receives the POSIX permissions of the POSIX group or everyone else field on the server - let's say there are no ACLs for the time being - and he becomes the owner of the copy. Since the original owner, root on the server, had read and write permissions, those permissions are preserved and given to tsmith for the copy.
    Another way to deal with this situation would be to use ACLs by placing an inheritable ACL granting full control or read and write for tsmith on his home folder. Any new items copied to or created in his home folder would inherit that ACL. Similarly, an inherited ACL deny entry (but not an explicitly-set entry) affecting tsmith and the delete and delete_child controls could be placed on the affected server share point's top-level before the original was copied there. Then, any subsequent copies from the server volume would drop this inherited ACL entry entirely. These would, however, be much more complicated situations, and your scenario is best handled in the POSIX-only fashion. (This doesn't mean that you have to disable ACLs on the server - just don't apply them for the share point!)
    --Gerrit
    search:Owner read only

  • How can i read my 1080 50p files from my panasonic camera

    Hi all newbie here can someone give me the name of software that can read my panosonic HDC-HS700 1080-50p files from my camera i can read the HA 1920 files okay but just cant read the 1080 file, people tell me that the imac doesnt support the 1080-50p files but surely there must be some software that can read from my camera.
                 Regards to all Pete.

    Hi  EZ  Jim, 
    After reading and following the info you sent through, i realised that the question i asked was totally wrong.
    my question is that my mac cant read or even see the 1080/50p on the hard drive of my cam, because apparently the Panasonic HDC-HS700 is not compatible with Apple can you beleive that, so now i have got to go out and by the software from Panasonic install on a crappy windows machine and then onto a portable hard drive and then back into my Mac.what a load of c--p.
    So thanks EZ for your help it was in any event usefull info that you gave me,
    My very Best regards Pete.
    EZ Jim wrote:
    You're welcome.
    We will watch for your status update.
    (Hope you and Fido have a nice walk! )
    Message was edited by: EZ Jim
    Mac OSX 10.8.3

Maybe you are looking for

  • Left Outer Join not working in Infoset.

    Hello Friends, I have two ODSes , one for planned data (zplan) and other is billing z ods (zsd_o03). now the situation is such that in my planned ODS  for one sold to party . SOLD TO        MATERRIAL CALMONT     PLANNED QUANTIY 14315            100  

  • XSAN No Controller

    Hi, My xsan shows no controller. Even I am unable to connect to Xsan. Even I the permission to Remove a computer is Grey But when I open from terminal using cvadmin it shows the controller. Can I add the previous controller again as its not showing i

  • Extractor for Table BKPF

    Hi ,   I wanted to know the extractor for table BKPF . Where can i know or which is the extratcor Regards Ankit Vaish

  • How to Change Document Default Status

    Hi Everyone, How can i change Document default status for Goods Receipt PO. I want like this When user create new GR PO the default status for the document should be "Draft" but now it shows status as "Open" Thanx & Regards, Amar

  • MacBook pro terminal gets  -bash-3.2 MacBook pro locked out

    get this when I go to try terminal on my MacBook pro. have tried just about every other way to getin to MacBook allhavefaileds, can I get in with terminal with proper command, and what is it, have tried some, but will not login