Count of no.of files passed for interface

Hi friends,
Can we have a Count of no.of files passed for interface per day..I need it to be an automated process like the no of files interface has processed has to be updated in the daily report.. Can we have it as automated other than manual check.. Is this possible?

Hi Mahesh,
Here are the tables where the message will be stored in tables.
Transaction Codes - (SE11, SE 16)
SXMSPMAS XI Messages - Master
SXMSPHIST XI Messages - Master History table
SXMSPEMAS Integration Engine: Enhanced Message Queue (Master)
SXMSPVERS Integration Engine: Message Version
SXMSPERROR XML Message Broker: Message Queue (Incorrect Entries)
SMPPMAP3 Mapping Runtime: Mapping
SMPPREL3 Mapping Runtime: Mapping Relation
SMPPSPLIT XI Mapping: Merge and Split
SXMSPFRAWH Integration Engine - Header Table for Raw Data for Performance Evaluation
SXMSPFRAWD Integration Engine: Data Table for Raw Data for Performance
Adapter Engine Tables -
Schema Name - SAPSR3DB
XI_AF_MSG AF XI Message
XI_AF_MSG_AUDIT audit log entries
If runing on PI 7.1, table names are as follows -
BC_MSG AF PI 7.1 Message
BC_MSG_AUDIT PI 7.1 audit log entries (**)
Thanks,

Similar Messages

  • Passing the file name and file type for mime_download

    HI All,
    Can You Please let me know how to pass the parameters file name and file type for mime_download.
    Thanks in Advance.
    Regards,
    Haritha.

    Hello Haritha,
    Please see the SAP help:
    http://help.sap.com/saphelp_nw70/helpdata/en/46/9bd97240b81e5be10000000a1553f6/frameset.htm
    Edgar

  • Servlet for reading lotus notes nsf-file via corba-interface

    Hello,
    we have written a servlet for reading lotus notes nsf-file via corba-interface.
    So we can read lotus notes-files in a portal url-iview.
    Are you interested?
    Best regards
    Oliver Prodinger

    If you check out the thread from my question (524480), we found
    that we had to enable the IIOP task on the domino server as we
    were not running it, this cured the problem I think you are
    encountering.
    This has not solved all our problems as I only see Notes data in
    the Inbox Portlet Preview mode, this problem is refered by other
    contributors as the collapsing Portlet problem (496625) and I do
    not know whether there is a solution to this problem yet.
    Hope this helps.

  • Generate XSL for File to mail interface

    HI there,
    Can somebody please tell me how to write xsl for the file to mail interface.
    I am getting only payload in the file and I want to convert that payload into xml so that I can map in PI.
    I tried with FCC but it doesn't work so the only option is via XSL mapping or JAVA mapping but unfortunately I don't know any of them, so can somebody please help me and write XSL for me if possible.
    Source File
    ACK3|24052010095354|4BC8612A7A530073E100800024590A61|1005241738SPBC8612A7A5000|D0000018|SXI|75130|(INFO) File Sent|PAYMENTS
    and I want the target file
    <ns0:ContentContainer xmlns:ns0="http://s.com/xi/10/lite">
       <payload>ACK3|24052010095354|4BC8612A7A530073E100800024590A61|1005241738SPBC8612A7A5000|D0000018|SXI|75130|(INFO) File Sent|PAYMENTS</payload>
    </ns0:ContentContainer>
    Thanks,
    Iqbal

    Hi Iqbal,
    The below mentioned do not seems to be difficult using FCC.
    In FCC mention following parameters. And provide required structure as message type in content conversion paramter.
    Recordset Structure :  row,*
    row.fieldNames:   payload
    row.fieldSeparator: 'nl'
    row.endSeparator: 'nl'

  • Error while testing File-PI-SOAP interface

    Dear All,
    We are working on File-PI-SOAP interface in which we have exposed a functional module from ECC server as receiver webservice. When we are testing the the interface we are getting following error :
    Delivering the message to the application using connection SOAP_http://sap.com/xi/XI/System failed, due to: com.sap.engine.interfaces.messaging.api.exception.MessagingException: java.io.IOException: invalid content type for SOAP: TEXT/PLAIN; HTTP 407 Proxy Authentication Required.
    In the receiver SOAP communication channel we have provided the required ECC login & proxy authentication credentials and also we are able to open the URL from PI server. Kindly let us know what more settings are required o make this interface running.
    Regards,
    NJ

    Hi,
    This is a synchronous interface and I want to send the output from webservice as a file to third party system.
    We are just working on a demo interface, so we have created a webservice of a function module of ECC and also our interface is File-PI-SOAP.
    We have also given the login credentials for proxy authentication but still we are getting the error the
    SOAP: error occured: com.sap.engine.interfaces.messaging.api.exception.MessagingException: java.io.IOException: invalid content type for SOAP: TEXT/PLAIN; HTTP 407 Proxy Authentication Required
    Kindly let us know, how can we resolve this error.
    Regards,
    NJ

  • Getting error while reading file from /data/interface

    HI Experts,
    while i am executing the job MPU* in ECC it write error file in /data/interface/sidclntxxx/IN/MPU742. now for the tcode ZMM_MPU_ANALYSIS it ask that error file, it is reflecting "error reading log file /data/interface...". now i am not getting the email also that supposed to get. please help
    Edited by: stiw382 on Dec 1, 2011 1:17 PM

    > while i am executing the job MPU* in ECC it write error file in /data/interface/sidclntxxx/IN/MPU742. now for the tcode ZMM_MPU_ANALYSIS it ask that error file, it is reflecting "error reading log file /data/interface...". now i am not getting the email also that supposed to get. please help
    is that file accessible on the operating system? Is the filesystem mounted?
    Z-transaction codes are customer developments, since nobody here knows, what this program does I suggest you contact the developer of that transaction.
    Markus

  • Programatically count the number of files in a folder

    Hi Folks
    I need to programatically count the number of files in a given folder and return the value to a Program Unit to be included in further processing.
    Can anyone point me in the right direction. I have looked at UTL_FILE but this does not offer what I want.
    Does anyone know of an obscure package in Oracle that I could use?
    I'm using 9i in a W2K environment.
    Many thanks
    Simon Gadd

    Hi,
    following a java based solution i found somewhere on Tom Kyte's AskTom forum.
    GRANT JAVAUSERPRIV to TESTUSER;
    dbms_java.grant_permission( 'TESTUSER', 'SYS:java.io.FilePermission', 'c:\', 'read' );
    create global temporary table DIR_LIST
    ( filename varchar2(255) )
    on commit delete rows
    create or replace and compile java source named "DirList"
    as
      import java.io.*;
      import java.sql.*;
      public class DirList
        public static void getList(String directory)
                           throws SQLException
          File path = new File( directory );
          String[] list = path.list();
          String element;
          for(int i = 0; i < list.length; i++)
              element = list;
    #sql { INSERT INTO DIR_LIST (FILENAME)
    VALUES (:element) };
    create or replace procedure get_dir_list( p_directory in varchar2 )
    as language java
    name 'DirList.getList( java.lang.String )';
    exec get_dir_list( 'c:\temp' );
    select count(*) from dir_list;

  • Lucreate - „Cannot make file systems for boot environment“

    Hello!
    I'm trying to use LiveUpgrade to upgrade one "my" Sparc servers from Solaris 10 U5 to Solaris 10 U6. To do that, I first installed the patches listed on [Infodoc 72099|http://sunsolve.sun.com/search/document.do?assetkey=1-9-72099-1] and then installed SUNWlucfg, SUNWlur and SUNWluufrom the S10U6 sparc DVD iso. I then did:
    --($ ~)-- time sudo env LC_ALL=C LANG=C PATH=/usr/bin:/bin:/sbin:/usr/sbin:$PATH lucreate -n S10U6_20081207  -m /:/dev/md/dsk/d200:ufs
    Discovering physical storage devices
    Discovering logical storage devices
    Cross referencing storage devices with boot environment configurations
    Determining types of file systems supported
    Validating file system requests
    Preparing logical storage devices
    Preparing physical storage devices
    Configuring physical storage devices
    Configuring logical storage devices
    Analyzing system configuration.
    Comparing source boot environment <d100> file systems with the file
    system(s) you specified for the new boot environment. Determining which
    file systems should be in the new boot environment.
    Updating boot environment description database on all BEs.
    Searching /dev for possible boot environment filesystem devices
    Updating system configuration files.
    The device </dev/dsk/c1t1d0s0> is not a root device for any boot environment; cannot get BE ID.
    Creating configuration for boot environment <S10U6_20081207>.
    Source boot environment is <d100>.
    Creating boot environment <S10U6_20081207>.
    Creating file systems on boot environment <S10U6_20081207>.
    Creating <ufs> file system for </> in zone <global> on </dev/md/dsk/d200>.
    Mounting file systems for boot environment <S10U6_20081207>.
    Calculating required sizes of file systems              for boot environment <S10U6_20081207>.
    ERROR: Cannot make file systems for boot environment <S10U6_20081207>.So the problem is:
    ERROR: Cannot make file systems for boot environment <S10U6_20081207>.
    Well - why's that?
    I can do a "newfs /dev/md/dsk/d200" just fine.
    When I try to remove the incomplete S10U6_20081207 BE, I get yet another error :(
    /bin/nawk: can't open file /etc/lu/ICF.2
    Quellcodezeilennummer 1
    Boot environment <S10U6_20081207> deleted.I get this error consistently (I ran the lucreate many times now).
    lucreate used to work fine, "once upon a time", when I brought the system from S10U4 to S10U5.
    Would anyone maybe have an idea about what's broken there?
    --($ ~)-- LC_ALL=C metastat
    d200: Mirror
        Submirror 0: d20
          State: Okay        
        Pass: 1
        Read option: roundrobin (default)
        Write option: parallel (default)
        Size: 31458321 blocks (15 GB)
    d20: Submirror of d200
        State: Okay        
        Size: 31458321 blocks (15 GB)
        Stripe 0:
            Device     Start Block  Dbase        State Reloc Hot Spare
            c1t1d0s0          0     No            Okay   Yes
    d100: Mirror
        Submirror 0: d10
          State: Okay        
        Pass: 1
        Read option: roundrobin (default)
        Write option: parallel (default)
        Size: 31458321 blocks (15 GB)
    d10: Submirror of d100
        State: Okay        
        Size: 31458321 blocks (15 GB)
        Stripe 0:
            Device     Start Block  Dbase        State Reloc Hot Spare
            c1t0d0s0          0     No            Okay   Yes
    d201: Mirror
        Submirror 0: d21
          State: Okay        
        Submirror 1: d11
          State: Okay        
        Pass: 1
        Read option: roundrobin (default)
        Write option: parallel (default)
        Size: 2097414 blocks (1.0 GB)
    d21: Submirror of d201
        State: Okay        
        Size: 2097414 blocks (1.0 GB)
        Stripe 0:
            Device     Start Block  Dbase        State Reloc Hot Spare
            c1t1d0s1          0     No            Okay   Yes
    d11: Submirror of d201
        State: Okay        
        Size: 2097414 blocks (1.0 GB)
        Stripe 0:
            Device     Start Block  Dbase        State Reloc Hot Spare
            c1t0d0s1          0     No            Okay   Yes
    hsp001: is empty
    Device Relocation Information:
    Device   Reloc  Device ID
    c1t1d0   Yes    id1,sd@THITACHI_DK32EJ-36NC_____434N5641
    c1t0d0   Yes    id1,sd@SSEAGATE_ST336607LSUN36G_3JA659W600007412LQFN
    --($ ~)-- /bin/df -k | grep md
    /dev/md/dsk/d100     15490539 10772770 4562864    71%    /Thanks,
    Michael

    Hello.
    (sys01)root# devfsadm -Cv
    (sys01)root# To be on the safe side, I even rebooted after having run devfsadm.
    --($ ~)-- sudo env LC_ALL=C LANG=C lustatus
    Boot Environment           Is       Active Active    Can    Copy     
    Name                       Complete Now    On Reboot Delete Status   
    d100                       yes      yes    yes       no     -        
    --($ ~)-- sudo env LC_ALL=C LANG=C lufslist d100
                   boot environment name: d100
                   This boot environment is currently active.
                   This boot environment will be active on next system boot.
    Filesystem              fstype    device size Mounted on          Mount Options
    /dev/md/dsk/d100        ufs       16106660352 /                   logging
    /dev/md/dsk/d201        swap       1073875968 -                   -In the rebooted system, I re-did the original lucreate:
    <code>--($ ~)-- time sudo env LC_ALL=C LANG=C PATH=/usr/bin:/bin:/sbin:/usr/sbin:$PATH lucreate -n S10U6_20081207 -m /:/dev/md/dsk/d200:ufs</code>
    Copying.
    *{color:#ff0000}Excellent! It now works!{color}*
    Thanks a lot,
    Michael

  • ERROR: MyService.jws:715:There are two or more operations with the same schema-element 'ns0:MyNameSpace' on the input message in a web service file or callback interface.

    I have two web service operations that have the same complex type as their input
    parameter. I want to map this type to an existing schema. I can successfully
    do this with the first operation using XQuery but when I attempt to do this with
    the second operation I get the following error:
    ERROR: MyService.jws:715:There are two or more operations with the same schema-element
    'ns0:MyNamespace' on the input message in a web service file or callback interface.
    ERROR:      SUGGESTION: Use different schema-element values for each of those operations.
    How can I use different schema-element values? The input parameters are to be
    mapped to the same schema and same element since they are the same for both operations....

    I am having the same problem. How did you resolve this..?? could you please tell me the solution??
    Thanks
    Shari

  • Limit of pages or file size for iBooks Author?

    iBooks Author v.1.1(190)
    I use iBooks Author for work, is there a maximum amount of pages iBooks can handle or limit to file size? I will have around 80+ pages including videos,links, and interactive widgets. My current page count is 52 and file size is 60MB, plus iBooks Author frequently quits when I save and preview on iPad 1.

    See Optimizing performance in your iBooks Author books [ support.apple.com/kb/HT5117 ] and know that max  book size is 2GB.
    60 mb shouldn't be an issue...

  • File-PI-SOAP Interface error

    Dear All,
    We are trying a interface in which we wanted to do Post Goods Issue in SAP from a file. We have exposed our Z-RFC which will do PGI as webservice and has imported its WSDL in PI.
    Then, we have created a File-PI-SOAP interface. Now, when we are testing the interface we are getting success message in SXMB_MONI but in RWB, message for picking the file is success and when I am opening the message details its showing
    " 500 Internal Server Error" ""Application error occurred during the request procession."
    Details: "The WebApplicationException log ID is [001A64D0A1FC005F0000004600002E4800CEA8DA2043BF9A]."
    Same error is showing for the receiver message but at payload I am able to see the correct values.
    Kindly let us know, if any configurations are to be required to remove thsi error.
    Regards,
    NJ

    Hi,
    >500 Internal Server Error
    does mean the server does not understand the request. I your case: You want to make use of a service which assumedly no exist. May be you have to release or activate it or to do something in SOAMANAGER. Please check your service first with SOAP UI.
    Regards,
    Udo

  • Why does OS X Finder recognize a folder as a file when it counts the number of files?

    When I select a folder in Finder and click cmd+i(or cmd+option+i), it shows the number of files in the folder. However, that number is weird for me, because they count a folder itself as a file. For example, when I have a single file in my Download folder and I clicked cmd+i on Download folder, then the number of files in a inspection window is 2, not 1. It is because that number includes not only the file inside the folder, but also the folder itself. (Check the below image)
    This is not a big problem if there is no hierarchical structure in my folder. But, if I have thousands of files and folders inside, it is really hard to figure out how many files(not folders) are inside. I cannot understand why Mac OS X counts a folder as a file. In fact, a folder is not an actual file, just a container.
    So, is there any philosophy to do like this? Or is there any way to count only the number of actual files?

    Sorry byron1st
    there is no way to count only files in a folder.
    Finder will count everything inside, folders and files.
    Information about a folder will count the same.
    If there are no subsubfolders in one folder, maybe a little trick can help:
    Select list view for the open folder, close all subfolders.
    Type command+# to see the footer of the window, that counts (hopefully) only the closed subfolders (content) in the open folder.
    Remember the amount.
    Hold option and click on one triangle in front of one subfolder. All subfolders will be opened!
    The difference between amount a and amount b is the number of files (and subsubfolders, if unfortunately inside…).

  • What is theoptimal file size for stills, and dpi?

    what is theoptimal file size for stills, and dpi?

    In video dpi is not relevent. What counts in video is pixels. In FCP most recommendations I have seen is to size your stills so they have no more than 3 times the pixels in your project aspect ratio. For example if your project is 1280x720 then stills should be no more than 3 times the 1280x720. This allows you to crop the stills and still have good resolution. If no cropping is anticipated then they could be the same size as your project.

  • Is there a simpler solution/directory/file structure for banner ads?

    I also have to wonder what a pain this has to be for a large website with many rotating banner ads - having all of these same files copied over and over in separate subdirectories. Flash had 2 files - a html file and swf file stored in the same directory. Am I missing something?
    the banner name - with config.xml - created by Dreamweaver
    Assets -  4 or 5 files, not counting json or sound files
    edge_includes  - and its 2 files
    images - at least they are unique
    And I, maybe wrongly, create a html file in Dreamweaver with the oam file inserted in it. 
    Bob

    From the Edge Animate JavaScript 2.0 API Docs:
    CDN Hosting
    Using the Adobe Content Distribution Network (CDN) is a great way to speed up Animate composition delivery. Compositions using the Adobe CDN all share a single URL for jQuery and the Edge Runtime. The browser caches the runtime, so the user only downloads the library once no matter how may Animate compositions they view, even if compositions are on different sites and produced by different authors.
    Don't use the CDN if your composition needs to run without an Internet connection or if you want to use your own hosting exclusively.
    The Edge Runtime CDN uses the following URL:
    <script type="text/javascript" charset="utf-8"    src="http://download.adobe.com/pub/adobe/edge/animate/2.0.0/edge.2.0.0.min.js"></script>
    So now I know what that CDN option is about.
    Bob

  • BPM-  ADF binding files architecture for ADF human taskflow.

    Hi,
    I wast to know what is the architecture behind the hundreds of the file generated for the BPM adf integration for the human taskflow.
    Is there any method of debugging why data coming to BPM human taskflow is not coming properly to ADF.
    I appeciate any debug pointers, although these files are clearly xml files and we have no control over generation of these files but still we want to know how data been populated and how if any error occured of data corruption, how we will resolve.
    REF we are facing major issues with the data corruption issues with the ADF BPM components.
    Thanks
    VIpin

    Hi Vipin
    This post has an interesting discussion related to your question. I am glad BPM is infact generating all these complex xml files behind the screens and taking care of so many things. Thanks to ADF framework. Yes, I totally agree, debugging may be nightmare to begin with. But with divide and rule, you can get over it. End of the Day its just pure Workflow and a complex ADF. ADF can be stripped out carefully and run all the actonis from a normal web application. If we design properly as pure SOA, we create back end services like EJBs that can be consumed from WebServices or ADF Layer. Once we have this we can integrate these Services into any BPM Process.
    Re: Integrating BPM taskflows into an ADF application
    Now, coming back to your questions:
    1. You have a Complex BPM Process with complex BO (business objects), Arrays of BOs. I hope you have XSD for your BOs and using them only you created BusinesObjects and ProcessDataObjects. Now lets assume you have a BPM Process that uses this complex data structure. Lets take a very first Human Task initiaor who may see a Task Details page with your compelx objects. Lets say, in Task Details page, he says a header section with 10 fields. Another section with a Table of Multiple Rows and columns (arrays of Data), etc etc.
    2. NOW the main question is, when this User Saves or Submits the Task, you should be doing something with that Data ? Am I right. Are you Saving it to any Database or everything is in the Payload itself. If so, I hope you have complex. If its payload itself, then I hope you have a complex strucutre to pass arrays of data through out the process.
    3. Assuming all this gets Stored into backend Database. Lets not worry about retrieving it back when task goes to second user. Always Payload will store the key IDs lke dataId like OrderId, CustomerId etc and using that it can pull any data from backend and show on task details page.
    4. This is where the ADF comes into picture. So even before you integrate your complex backend data with Process, first create a simple WebAppilcation. If you have EJB methods to handle this data, you create Data controls, bindings. Bindings will help generate the Screens. So you can first test your application completely here. Once all this is working, thats when you call these Services from your Process.
    And beauty is with new release of BPM Suite (should be in BPM 11.5 + FP (Feature Pack)), they have a concept of Simulation. This should help you to simulate the Process without any real data and JUST focus and test on the Flow of the Process. This comes real handy for testing.
    And for the data used in your process, you have ADF Tools to create stand alone ADF Application.
    We have some applications like yours with very complex data structure, arrays that gets stored like in 25 tables with references going everywhere. First we built a simple data layer using EJBs and JPA layer (JPA Entities). These Entities have annotated methods to take care of most of the things. In the EJB we exposed some generic methods instead of adf generating them, due to complex data structre. Then we created data controls and bindings. Create Custom JSPX Files. This is where we ended up having our own Layout for the Data. I hope you have already seen by default, task details layout page generated is NOT usabe at all. It generates all the data in one Single Long column. We do not use that. Since we know data, we created our own Task details page. The only thing we add is the Actions bar and comments and attachment section. When you auto generate a taskdetails page for any .TASK file, you will see DataBindins created for Payload, Actions, and also separately for coments and attachmetns. Just drag these comments and attachments on to your page and define your own layout.
    Re: deploying a large Oracle BPM Application with multiple UI projects (Single taskdetails for multiple UIs and Tasks)
    Re: What exactly 'PAYLOAD' Means??? (Simple definition for Payload word referred in any Process)
    If you see one of the old posts, there is a way to reuse just one Single TaskDetals page for Multiple different Human .TASK files. All that matters iis the top Actions Menu that are differenet for each human task like someone can Submit, someone can only Approve or Reject etc. We can club all these Actions into single actionsMenu.jsff and put this on TaskDetails page and assing to all the .task files. Menus are automatically disabled/enabled using workflow APIs. You do not need any special coding for them. Look at code snipet for any actions button that has like isAllowedAction and pass the name of the action. Workflow APIs see if that .TASK file has a Action named with that internally.
    In conclusion, there is a decent clear gap between Process and the ADF Generated Task Details page.
    Thanks
    Ravi Jegga

Maybe you are looking for