FEDEX / UPS  real-time interface

Hello All,
did any one worked with real-time interface to FEDEX/UPS using XI ?
any input greatly appreciated?
Thanks

Ravi
My company specializes on multi carrier shipping solutions. We have an interfacing white paper if you are interested.
Before using the interfaces described above you need to consider the following attributes about your shipping. The thing is the carriers systems have both legal licensing restrictions, performance issues and compliance issues that you need to consider.
- Do you need to rate compare between carriers. ( This is a legal issue with the on-line services ).
- What modes of transport are you going to ship. Is it all parcel or are there some freight ( not all the services are available through the API's ).
- What volume? ( this is key.. the on-line carrier API's are designed for a couple of hundred parcels per day max ).
- How often does you company change carriers? ( each carrier is constantly changing and there is no interface standard in the industry ).
If you want to chat more please feel free to call me on 508.787.1842. I would also be happy to e-mail you interfacing white papers if thats of interest. [email protected]
Good Luck
Matthew

Similar Messages

  • How much interfaces can be driven in "real-time" with a VI in LABVIEW 6i

    In my application I need to merged data from different devices hanging to my computer via different interface types. Each data set in to be specifically formatted in a file. Each data set comes a a rate of ~ 10 Hz.
    As acquisition devices I do have
    - 4 digitizers connected to the PCI bus and their labview driver are already given by the constructor
    - an instrument from which data to be imported via a GPIB interface
    - an instrument from which data are imported via a serial port
    - and probably in the future, 2 devices mounted on two USB port
    I wish to developp a LABVIEW 6i code which permit me to make the request to all instrument, to import the data and to merged them as quick as possible so
    that I can store data set coming every 0,1 s. All data from all instruments must be recorded for the same event at the same time (in a triggered manner) before it performs the acquistions for the next event
    I would like to know what is the best way to each my goal.
    - Is it possible to control all this with one VI (containing lot of VIs) ?
    - Should I execute a VI for each interface and run them at the same time and save all data set with a time stamps in a file and then, off-line, process all the files and merge my information.
    In a general manner, are there any rules which indicate the maximum number of interface (according to their nature) which can be properly controled on a given platform from one computer.
    Thanks you very much for any answers

    Its been a couple of days with no reply....
    Look at the occurence functions.
    You can have one loop send occurences at teh rate you need.
    Put each of the other functions in seperate loops that wait for the occurence on each interation.
    Provided each of your loops runs fast enough this approach should serve you up till you run into the indeterminism of Windows. If the the indeterminism will be an issue, you may have to go with LV Real-Time.
    Ben
    Ben Rayner
    I am currently active on.. MainStream Preppers
    Rayner's Ridge is under construction

  • ABAP interface in real time

    Dear all,
    Hereafter is my scenario;
    My customer after sales system works on AS400. Assuming that I would manage spare parts (required for after sales) inventory, including goods entry, goods issue as well as the company financial accounting in SAP, I have to interface the two system.
    From SAP to AS400: (flow1) Master files + (flow2) goods entry
    From AS400 to SAP: (flow3) AR postings, (flow4) goods consumption
    Then my two systems will be in line in terms of referential data (flow1), stock movement (flow2 + flow4), invoicing (flow3).
    1) My question is to know what would be the best way to flow2 knowing that stock movement following a goods entry (done in SAP) should be visible in real time from AS400?
    2) Regarding other flows, no need for real time. So I guess flat file should be ok. However, would master data for instance sent from SAP to AS400 require a mapping table in order to be properly integrated in AS400?
    Regards.

    Real Time doesn't mean always fast. It depends on your application requirements.
    Anyway, don't think too much about Real Time in Windows. If it's possible, try to do all "time critical" operations in Matlab and after that pull data with LV.

  • FieldPoint real-time controller or Ethernet interface

    Good Afternoon,
    The first feature listed on the FP-2000 product page is:
    Stand-alone embedded real-time controller or Ethernet interface for PC-based distributed I/O.
    Has anybody successfully implented a FP-2000 as an Ethernet interface for PC-based distributed I/O?  In other words, have the FP-2000 act like a FP-1600?
    There are some discussion forum threads that dance around the subject but I have not found one that really answers it.
    Thanks,
    Ed
    Solved!
    Go to Solution.

    Hi edlad,
    After playing around with a FP-2010 for a while, I do believe that you need to install LabVIEW Real-Time on the controller in order for it to work, even if you are technically not using it because you are not running a VI/executable on the controller. As far as your issue with installing LabVIEW 8.5.1 on a FP-2000, that is troubling because there should not be an issue doing this at all. In fact, I have personally been able to install LabVIEW 8.5.1 on many FP-2000s. For that issue, I would suggest reformatting the FP-2000 from MAX and try again.
    ThinkG: Regarding your question about using a cFP-2220 to communicate with Lookout, I am not sure because Lookout support is handled by our Shanghai branch via email only. However, I do know that you can use that second ethernet port to communicate with another FieldPoint network module (cFP-180x) using the FieldPoint drivers. Here are a couple good KnowledgeBase articles regarding that configuration. I hope they are helpful.
    http://digital.ni.com/public.nsf/allkb/F602F6F1B243282686257495007695BB?OpenDocument
    http://digital.ni.com/public.nsf/allkb/67F94BB93BCE32CF86257367006B3659?OpenDocument
    Thank you for choosing National Instruments.
    Aaron Pena
    National Instruments
    Applications Engineer
    http://www.ni.com/support 

  • Implementing Real Time, Outbound Interfaces

    Hi All,
    I have never worked on Interfaces before.
    I need help in Real time, Outbound Interfaces.
    How do we implement Real Time and Outbound Interfaces.
    Any pointers or suggestions would really be helpful.
    Thanks,
    Saher

    cotton.m wrote:
    JMS? Thanks cotton, I will look into JMS
    Unless you just wanted to play with sockets for the experience in which case you can do as you like. If you want get better with Java networking then IMO you must get ejp's book. [http://www.amazon.com/Fundamental-Networking-Java-Esmond-Pitt/dp/1846280303]
    It's more of a fix before we release it to the poor souls that will be using our software, oh how the public are blind...
    Mel

  • How do you interface a Renishaw laser real-time with NI DAQ?

    I understand that it is possible to acquire "real time" feedback from a Rensishaw Laser while measuring linear position, with the following modifiers applied.
    *The standard output is a digitized value to a proprietary (Renishaw) data acquisition (DAQ) card.
    *Renishaw lasers have a quadrature output facility on the back of the laser head. This permits capture of the �raw� interferometry signal. This raw signal is not compensated for environmental factors such as temperature, air pressure, humidity, etc.
    *The standard system integrates signals from an environmental unit for air temperature, pressure, and humidity, as well as signals from up to three material probes. Acquisition and pro
    cessing of these or equivalent signals would have to be considered to provide data for environmental compensation calculations (Edlen equation). This is mandatory in order to obtain accurate distance and velocity readings.
    *The quadrature output can communicate with other cards. (NI DAQ?)
    Specifics with regard to Labview software and required DAQ to read real time Laser output would be very much appreciated.

    You can use a counter available on the E-series MIO DAQ boards or the 660x counter boards. You can measure the position, frequency, and period of the output from a quadrature encoder. Examples on how to measure signals from a quadrature encoder for LabVIEW can be found on the web at ww.ni.com/support under the technical resource section search under the example programs section.
    You may also want to contact Renishaw to see if they have drivers available for LabVIEW and their DAQ equipment.

  • How do I generate a compiled app with a user interface running real-time?

    Hi everyone,
    I've developed a simple application that uses a PXI chassis to run.  It toggles some discretes and does some simple things, but it doesn't really have any "real-time" requirements.  We're just using the PXI chassis because we have it and we need the discrete I/O.
    My question is this - how can I convert this application to a compiled real-time app and yet maintain the front panel on the local PC.  LabView appears to simply do this for me.  I am connected to the real-time target via ethernet.  When I compiled the program, it puts it in the target's startup directory, but how do I run the front panel from the local PC?
    Thanks,
    Jason

    Hi Jason,
    Below are some links that should get you started with deploying an application onto your real-time target and then controlling it from a PC that does not have the LabVIEW Development environment:
    http://digital.ni.com/public.nsf/websearch/C90410C685CFF89786256C24005AC027?OpenDocument
    Launching code on the RT target:
    http://zone.ni.com/devzone/conceptd.nsf/webmain/7D7CFAFF379531B086256AC70058812B?opendocument&node=1...
    Deploying and launching a real-time app:
    http://zone.ni.com/devzone/conceptd.nsf/webmain/9ac4955881bd7b4d86256d0b0061dd1c
    Application Builder and RT PXI controller:
    http://zone.ni.com/devzone/conceptd.nsf/webmain/81b6674f3556599b862568cf005d26c6
    RT good programming practices (use RT communication wizard for easy implementation):
    http://zone.ni.com/devzone/devzoneweb.nsf/Opendoc?openagent&E4C4DCD9C00295A186256B5100665BF5
    RT Communication Wizard:
    http://zone.ni.com/devzone/conceptd.nsf/webmain/a6f17ee4adcab99686256d5e0053e210
    I would suggest going through these tutorials and documents, they should provide you with what you need to get started.  Please don't hesitate to ask any questions that you have a long the way.
    Bryan Snarr
    Field Engineer
    Northern California
    National Instruments

  • FEDEX, UPS, DHL and Pitney Bowes

    Is there any information on interfaces with the FEDEX, UPS and DHL with SAP for getting the tracking number for shipments? Also where can I find information on integration with Pitney Bowes for sorting and mailing solution with SAP?
    I am looking for direct interface options without any middleware. There is an OSS note 527342 which describes Fedex interface, however, the links in the notes are outdated and not working.
    Regards

    Hello,
    There is a WebEx on 12/13/2006 & 12/15/2006 @ 8:00 AM PST (California Time Zone).
    Agenda:
    Enterprise Centralized Shipping (ECS) For SAP With out middleware.
    Description: of ECS: ECS eliminates Black-box systems. Eliminates interfaces. Processweaver Solutions are developed based on ABAP and Netweaver platform.
    DATE: Wednesday, December 13, 2006
    TIME: 8:00 am, Pacific Standard Time (GMT -08:00, San Francisco ) .
    MEETING NUMBER: 804 681 562
    PASSWORD: weaver3
    Call-in number (US/Canada): 650-429-3300
    https://processweaver.webex.com/processweaver/j.php?ED=91808247&UID=52406287
    More information: email to [email protected]
    This is a great WebEx session to know more know more about SAP Enterprise Centralized Shipping (ECS)
    Thank You,

  • Questions on working in real time

    Hi.  Just have a couple of questions about working in real time, actually on the swf.
    Firstly, I want to create a sort of drag and drop interface.  Is it possible to have one stage act as two seperate stages?  By this, I mean is it possible to have the left hand side of my stage hold all the components, they get dragged to the right hand side of the stage and placed anywhere, the user can then click save, and only the right hand side of the stage is saved?
    Secondly, I am testing a few things out.  What would I do to get a type of free transform on an image on the stage in real time, in the swf?  So once the movie is played, the user could then resize any image how they like by dragging a corner of the image..
    Cheers for any advise.

    What I would like is the movie to display to the user.  When they see the movie, it looks like the screen is split in two.  On the left hand side of the screen will be the components which can be dragged and dropped.  The right hand side of the screen will be a birthday card, whereby the users can drop and drag components to personalise the card.  I then want the user to be able to save the card, but a normal save will save the whole screen, including the components side of the screen.  All I want is the right hand side of the screen saved, where their card has been created.  Would this be done with two movies, and communication between the two.  Or is this possible to do on just one stage?

  • Error -1074384569; NI-XNET: (Hex 0xBFF63147) The database information on the real-time system has been created with an older NI-XNET version. This version is no longer supported. To correct this error, re-deploy your database to the real-time system.

    Hello
    I have a VeriStand-Project (VSP) created with my Laptop-Host (LTH) which works with my PXI, while
    deploying it from my LTH. Then I have installed the whole NI enviroment for PXI and VeriStand use on a
    industrial PC (iPC). I have tried to deploy my VSP from the iPC to the PXI but the following error
    message arose on my iPC:
    The VeriStand Gateway encountered an error while deploying the System Definition file.
    Details: Error -1074384569 occurred at Project Window.lvlibroject Window.vi >> Project
    Window.lvlib:Command Loop.vi >> NI_VS Workspace ExecutionAPI.lvlib:NI VeriStand - Connect to System.vi
    Possible reason(s):
    NI-XNET:  (Hex 0xBFF63147) The database information on the real-time system has been created with an
    older NI-XNET version. This version is no longer supported. To correct this error, re-deploy your
    database to the real-time system. ========================= NI VeriStand:  NI VeriStand
    Engine.lvlib:VeriStand Engine Wrapper (RT).vi >> NI VeriStand Engine.lvlib:VeriStand Engine.vi >> NI
    VeriStand Engine.lvlib:VeriStand Engine State Machine.vi >> NI VeriStand Engine.lvlib:Initialize
    Inline Custom Devices.vi >> Custom Devices Storage.lvlib:Initialize Device (HW Interface).vi
    * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * • Unloading System
    Definition file... • Connection with target Controller has been lost.
    The software versions of the NI products (MAX/My System/Software) between my LTH and the iPC are
    almost the same. The only differences are:
    1. LabView Run-Time 2009 SP1 (64-bit); is installed on LTH but missing on iPC. The iPC has a 32-bit system.
    2. LabView Run-Time 2012 f3; is installed on LTH but missing on iPC.
    3. NI-DAQmx ADE Support 9.3.5; something strage on the LTH, because normally I am using NI-DAQmx 9.5.5 and all other DAQmx products on my LTH are 9.5.5. That means NI-DAQmx Device Driver 9.5.5 and NI-DAQmx Configuration 9.5.5.. On the iPC side all three products are 9.5.5.. That means NI-DAQmx ADE Support 9.5.5, NI-DAQmx Device Driver 9.5.5 and NI-DAQmx Configuration 9.5.5..
    4. Traditional NI-DAQ 7.4.4; The iPC has this SW installed. On the LTH this SW is missing.
    In order to fix this problem I have formatted my PXI and I have installed the following SW from the iPC:
    1. LabVIEW Real-Time 11.0.1
    2. NI-488.2 RT 3.0.0
    3. NI_CAN 2.7.3
    Unfortunately the above stated problem still arose.
    What can I do to fix this problem?
    I found a hint on http://www.labviewforum.de/Thread-XNET-CAN-die-ersten-Gehversuche.
    There it is written to deploy the dbc file againt.
    If this is a good hint, so how do I deploy a dbc file?
    I would feel very pleased if somebody could help me! :-)
    Best regards
    Lukas Nowak

    Hi Lukas,
    I think the problem is caused by differenet drivers for the CAN communication.
    NI provides two driver for CAN: NI-CAN and NI-XNET.
    NI-CAN is the outdated driver which is not longer used by new hardware. NI replaced the NI-CAN driver with NI-XNET some years ago, which supports CAN, LIN and the FLEXRAY communication protocol.
    You wrote:
    In order to fix this problem I have formatted my PXI and I have installed the following SW from the iPC:
    3. NI_CAN 2.7.3
    NI CAN is the outdated driver. I think that you should try to install NI-XNET instead of NI-CAN on your PXI-System, to get rid of the error message.
    Regards, stephan

  • BAdI........How are they used in the real time scenario.

    Im a rookie in ABAP development and i was wondering that how we use the BAdI enhancement technique in the real time scenarios. Like do we get create them or we have few already built BAdI's which we work on? Can any body please explain how this works.

    Hi Ramana,
    Business add-ins are enhancements to the standard version of the system.
    Business Add-In is a new SAP enhancement technique based on ABAP Objects.
    They can be inserted into the SAP system based on specific user requirements.
    Each Business Add-In has:
    • at least one Business Add-In definition
    • a Business Add-In interface
    • a Business Add-In class that implements the interface
    In order to enhance a program, a Business Add-In must first be defined
    Subsequently two classes are automatically generated:
    • An interface with ‘IF_EX_’ inserted between the first and second characters of the BADI name.
    • An adapter class with ‘CL_EX_’ inserted between the first and second characters of the BADI name.
    The Application developer creates an interface for this Add-In.
    There are multiple ways of searching for BADI.
    • Finding BADI Using CL_EXITHANDLER=>GET_INSTANCE
    • Finding BADI Using SQL Trace (TCODE-ST05).
    • Finding BADI Using Repository Information System (TCODE- SE84).
    1. Go to the Transaction, for which we want to find the BADI, take the example of Transaction VD02. Click on System->Status. Double click on the program name. Once inside the program search for ‘CL_EXITHANDLER=>GET_INSTANCE’.
    Make sure the radio button “In main program” is checked. A list of all the programs with call to the BADI’s will be listed.
    The export parameter ‘EXIT_NAME’ for the method GET_INSTANCE of class CL_EXITHANDLER will have the user exit assigned to it. The changing parameter ‘INSTANCE’ will have the interface assigned to it. Double click on the method to enter the source code.Definition of Instance would give you the Interface name.
    2. Start transaction ST05 (Performance Analysis).
    Set flag field "Buffer trace"
    Remark: We need to trace also the buffer calls, because BADI database tables are buffered. (Especially view V_EXT_IMP and V_EXT_ACT)
    Push the button "Activate Trace". Start transaction VA02 in a new GUI session. Go back to the Performance trace session.
    Push the button "Deactivate Trace".
    Push the button "Display Trace".
    The popup screen "Set Restrictions for Displaying Trace" appears.
    Now, filter the trace on Objects:
    • V_EXT_IMP
    • V_EXT_ACT
    Push button "Multiple selections" button behind field Objects
    Fill: V_EXT_IMP and V_EXT_ACT
    All the interface class names of view V_EXT_IMP start with IF_EX_. This is the standard SAP prefix for BADI class interfaces. The BADI name is after the IF_EX_.
    So the BADI name of IF_EX_CUSTOMER_ADD_DATA is CUSTOMER_ADD_DATA
    3. Go to “Maintain Transaction” (TCODE- SE93).
    Enter the Transaction VD02 for which you want to find BADI.
    Click on the Display push buttons.
    Get the Package Name. (Package VS in this case)
    Go to TCode: SE84->Enhancements->Business Add-inns->Definition
    Enter the Package Name and Execute.
    Here you get a list of all the Enhancement BADI’s for the given package MB.
    Have a look at http://help.sap.com/saphelp_nw04/helpdata/en/04/f3683c05ea4464e10000000a114084/content.htm
    http://help.sap.com/saphelp_erp2005/helpdata/en/73/7e7941601b1d09e10000000a155106/frameset.htm
    http://support.sas.com/rnd/papers/sugi30/SAP.ppt
    http://www.sts.tu-harburg.de/teaching/sap_r3/ABAP4/abapindx.htm
    http://members.aol.com/_ht_a/skarkada/sap/
    http://www.ct-software.com/reportpool_frame.htm
    http://www.saphelp.com/SAP_Technical.htm
    http://www.kabai.com/abaps/q.htm
    http://www.guidancetech.com/people/holland/sap/abap/
    http://www.planetsap.com/download_abap_programs.htm
    http://help.sap.com/saphelp_nw04/helpdata/en/c8/1975cc43b111d1896f0000e8322d00/content.htm
    /people/thomas.weiss/blog/2006/04/03/how-to-define-a-new-badi-within-the-enhancement-framework--part-3-of-the-series
    /people/thomas.weiss/blog/2006/04/18/how-to-implement-a-badi-and-how-to-use-a-filter--part-4-of-the-series-on-the-new-enhancement-framework
    How to develop BADI
    Reward if useful.
    Thanks
    Aneesh.

  • Help needed with CVI real time

    Hi,
    I am new to CVI real time and need help related to this. I will explain my scenario:
    1. I have few configuration files (INI). Need to validate the files (range checking..etc) during start of the application.
    2. When user clicks "START" in the user interface on the host machine, I need to generate outputs (analogs/discretes...) based on the settings in the config file and read some inputs from other instruments. ( I/O tasks like generating outputs/reading inputs i am planning to perform in RT side)
    3. From the user interface,  user can also change the config files. if user changes the files, again file validation needs to be performed. After changing the files, if user clicks START, need to take the newly enetered config files to perfom I/O tasks.
    4. I am NOT using reflective memory for my application.
    I am confused in the following area:
    1. File validation during startup , do i have to perform on both host and RT side?????
    2. if configuration files are changed in the host side by the user, how i have to send this latest file names to RT??? I think i should not send file names, i need to read from the file and its contents i need to pass to RT....Pls correct me , I am not sure about this....
    3. If I have to send file contents to the RT, how I have to do that....means i have to use structures,..???
    Please guide me. Any help would be highly appreciated.
    Regards,
    haari
    Solved!
    Go to Solution.

    Hey haari,
    1. File validation during startup , do I have to perform on both host and RT side?????
    This depends on the full range of responsibility for the INI files. If they are all needed simply to change what the I/O tasks are for, then you would technically only need to validate the INI file contents on the RT side. You could, however, validate on both sides if you wanted to. However, I would probably set it up such that I send a validation command to the target, have the target run a component of the code that validates the INI contents based on how I instructed it, and then return a message that says whether or not it met the requirements I specified. However, if you are concerned with offloading that process to the user desktop (instead weighing down the RT controller performance), you could FTP the file off of the controller and perform the validation process on the host side.
    2. if configuration files are changed in the host side by the user, how i have to send this latest file names to RT??? I think i should not send file names, i need to read from the file and its contents i need to pass to RT....Pls correct me , I am not sure about this....
    You can send them to the target over FTP. This can be done programmatically, through MAX, command window, or a Web browser.
    3. If I have to send file contents to the RT, how I have to do that....means i have to use structures,..???
    This is basically addressed in answer to question two. You would not need another structure in your RT code to handle the FTP receive component as this is handled by the FTP Client/Server communication. However, in the event that you successfully FTP a file to the target, you should likely send a message from the host to the target so that the target can respond accordingly and reload the file, reestablish the I/O values, run an idle procedure, or do whatever you like.
    Hope this helps. Have a great day!
    Tim A.
    National Instruments

  • What is the best way to store data on a network hard drive using CompactRIO RTOS and Labvew Real time?

    HI!
    I'm starting a project in which I have a low rate stream of data to read in a real time environment. I should store these data on a network hard disk without any PC with standard OS, I just have CompactRIO RTOS. How can I send this data to the network drive? Is it possible to just “write” data like I do for a standard file in LabView?
    Thanks for any help!!
    Il Conte
    dr. Valentino Tontodonato

    Il Conte,
    you have to keep in mind that normally the RT OS does not map drives other than the Compact Flash that it has onboard (C:\). There are exceptions such as
    -cFP-20xx which may have additional Flash Drives which can be addressed as D:\ Drive
    -CVS systems with IEEE-1394 interface can write/read to Firewire external Harddrives
    -PXI Controllers booted from a Floppy disk may map the floppy drive as A:\
    One solution to your needings may be to write data to files locally on your onboard CompactFlash and then transfer these files to a network location using FTP, provided the network drive you are pointing to supports FTP.
    Let us know if you need any more help with this,
    AlessioD
    National Instruments

  • Abap-hr real time questions

    hi friends
    kindly send me ABAP-HR REAL TIME QUESTION to my mail [email protected]
    Thanks&Regards
    babasish

    Hi
    Logical database
    A logical database is a special ABAP/4 program which combines the contents of certain database tables. Using logical databases facilitates the process of reading database tables.
    HR Logical Database is PNP
    Main Functions of the logical database PNP:
    Standard Selection screen
    Data Retrieval
    Authorization check 
    To use logical database PNP in your program, specify in your program attributes.
    Standard Selection Screen
    Date selection
    Date selection delimits the time period for which data is evaluated. GET PERNR retrieves all records of the relevant infotypes from the database.  When you enter a date selection period, the PROVIDE loop retrieves the infotype records whose validity period overlaps with at least one day of this period.
    Person selection
    Person selection is the 'true' selection of choosing a group of employees for whom the report is to run.
    Sorting Data
    · The standard sort sequence lists personnel numbers in ascending order.
    · SORT function allows you to sort the report data otherwise. All the sorting fields are from infotype 0001.
    Report Class
    · You can suppress input fields which are not used on the selection screen by assigning a report class to your program.
    · If SAP standard delivered report classes do not satisfy your requirements, you can create your own report class through the IMG.
    Data Retrieval from LDB
    1. Create data structures for infotypes.
        INFOTYPES: 0001, "ORG ASSIGNMENT
                            0002, "PERSONAL DATA
                            0008. "BASIC PAY
    2. Fill data structures with the infotype records.
        Start-of-selection.
             GET PERNR.
        End-0f-selection. 
        Read Master Data
    Infotype structures (after GET PERNR) are internal tables loaded with data.
    The infotype records (selected within the period) are processed sequentially by the PROVIDE - ENDPROVIDE loop.
              GET PERNR.
                 PROVIDE * FROM Pnnnn BETWEEN PN/BEGDA AND PN/ENDDA
                        If Pnnnn-XXXX = ' '. write:/ Pnnnn-XXXX. endif.
                 ENDPROVIDE.
    Period-Related Data
    All infotype records are time stamped.
    IT0006 (Address infotype)
    01/01/1990   12/31/9999  present
              Which record to be read depends on the date selection period specified on the
              selection screen. PN/BEGDA PN/ENDDA.
    Current Data
    IT0006 Address  -  01/01/1990 12/31/9999   present
    RP-PROVIDE-FROM-LAST retrieves the record which is valid in the data selection period.
    For example, pn/begda = '19990931'    pn/endda = '99991231'
    IT0006 subtype 1 is resident address
    RP-PROVIDE-FROM-LAST P0006 1 PN/BEGDA PN/ENDDA.
    Process Infotypes
    RMAC Modules - RMAC module as referred to Macro, is a special construct of ABAP/4 codes. Normally, the program code of these modules is stored in table 'TRMAC'. The table key combines the program code under a given name. It can also be defined in programs.The RMAC defined in the TRMAC can be used in all Reports. When an RMAC is changed, the report has to be regenerated manually to reflect the change.
    Reading Infotypes - by using RMAC (macro) RP-READ-INFOTYPE
              REPORT ZHR00001.
              INFOTYPE: 0002.
              PARAMETERS: PERNR LIKE P0002-PERNR.
              RP-READ-INFOTYPE PERNR 0002 P0002 .
              PROVIDE * FROM P0002
                  if ... then ...endif.
              ENDPROVIDE.
    Changing Infotypes - by using RMAC (macro) RP-READ-INFOTYPE. 
    · Three steps are involved in changing infotypes:
    1. Select the infotype records to be changed;
    2. Make the required changes and store the records in an alternative table;
    3. Save this table to the database;
    The RP-UPDATE macro updates the database. The parameters of this macro are the OLD internal table containing the unchanged records and the NEW internal table containing the changed records. You cannot create or delete data. Only modification is possible.
    INFOTYPES: Pnnnn NAME OLD,
    Pnnnn NAME NEW.
    GET PERNR.
        PROVIDE * FROM OLD
               WHERE .... = ... "Change old record
               *Save old record in alternate table
               NEW = OLD.
        ENDPROVIDE.
        RP-UPDATE OLD NEW. "Update changed record
    Infotype with repeat structures
    · How to identify repeat structures.
    a. On infotype entry screen, data is entered in table form.
        IT0005, IT0008, IT0041, etc.
    b. In the infotype structure, fields are grouped by the same name followed by sequence number.
        P0005-UARnn P0005-UANnn P0005-UBEnn
        P0005-UENnn P0005-UABnn
    Repeat Structures
    · Data is entered on the infotype screen in table format but stored on the database in a linear  
      structure.
    · Each row of the table is stored in the same record on the database.
    · When evaluating a repeat structure, you must define the starting point, the increment and the
      work area which contains the complete field group definition.
    Repeat Structures Evaluation (I)
    · To evaluate the repeat structures
       a. Define work area.
           The work area is a field string. Its structure is identical to that of the field group.
       b. Use a DO LOOP to divide the repeat structure into segments and make it available for  
           processing in the work area, one field group (block) at a time.
    Repeat Structures Evaluation(II)
    Define work area
    DATA: BEGIN OF VACATION,
                  UAR LIKE P0005-UAR01, "Leave type
                  UAN LIKE P0005-UAN01, "Leave entitlement
                  UBE LIKE P0005-UBE01, "Start date
                  UEN LIKE P0005-UEN01, "End date
                  UAB LIKE P0005-UAB01, "Leave accounted
               END OF VACATION.
    GET PERNR.
         RP-PROVIDE-FROM-LAST P0005 SPACE PN/BEGDA PN/ENDDA.
         DO 6 TIMES VARYING VACATION
                 FROM P0005-UAR01 "Starting point
                     NEXT P0005-UAR02. "Increment
                 If p0005-xyz then ... endif.
          ENDDO.
    Processing 'Time Data'.
    · Dependence of time data on validity period
    · Importing time data
    · Processing time data using internal tables
    Time Data and Validity Period
    · Time data always applies to a specific validity period.
    · The validity periods of different types of time data are not always the same as the date selection period specified in the selection screen.
    Date selection period |----
    |
    Leave |----
    |
    · PROVIDE in this case is therefore not used for time infotypes.
    Importing Time Data
    · GET PERNR reads all time infotypes from the lowest to highest system data, not only those within the date selection period.
    · To prevent memory overload, add MODE N to the infotype declaration. This prevents the logical database from importing all data into infotype tables at GET PERNR.
    · Use macro RP-READ-ALL-TIME-ITY to fill infotype table.
    INFOTYPES: 2001 MODE N.
    GET PERNR.
        RP-READ-ALL-TIME-ITY PN/BEGDA PN/ENDDA.
        LOOP AT P0021.
             If P0021-XYZ = ' '. A=B. Endif.
        ENDLOOP.
    Processing Time Data
    · Once data is imported into infotype tables, you can use an internal table to process the interested data.
    DATA: BEGIN OF ITAB OCCURS 0,
                  BUKRS LIKE P0001-BUKRS, "COMPANY
                  WERKS LIKE P0001-WERKS, "PERSONNEL AREA
                  AWART LIKE P2001-AWART, "ABS./ATTEND. TYPE
                  ASWTG LIKE P2001-ASWTG, "ABS./ATTEND. DAYS
               END OF ITAB.
    GET PERNR.
    RP-PROVIDE-FROM-LAST P0001 SAPCE PN/BEGDA PN/ENDDA.
    CLEAR ITAB.
    ITAB-BUKRS = P0001-BURKS. ITAB-WERKS = P0001-WERKS.
    RP-READ-ALL-TIME-ITY PN/BEGDA PN/ENDDA.
    LOOP AT P2001.
          ITAB-AWART = P2001-AWART. ITAB-ASWTG = P2001-ASWTG.
          COLLECT ITAB. (OR: APPEND ITAB.)
    ENDLOOP.
    Database Tables in HR
    ·  Personnel Administration (PA) - master and time data infotype tables (transparent tables).
       PAnnnn: e.g. PA0001 for infotype 0001
    ·  Personnel Development (PD) - Org Unit, Job, Position, etc. (transparent tables).
       HRPnnnn: e.g. HRP1000 for infotype 1000
    ·  Time/Travel expense/Payroll/Applicant Tracking data/HR work areas/Documents (cluster  
       PCLn: e.g. PCL2 for time/payroll results.
    Cluster Table
    · Cluster tables combine the data from several tables with identical (or almost identical) keys
      into one physical record on the database.
    . Data is written to a database in compressed form.
    · Retrieval of data is very fast if the primary key is known.
    · Cluster tables are defined in the data dictionary as transparent tables.
    · External programs can NOT interpret the data in a cluster table.
    · Special language elements EXPORT TO DATABASE, IMPORT TO DATABASE and DELETE
      FROM DATABASE are used to process data in the cluster tables.
    PCL1 - Database for HR work area;
    PCL2 - Accounting Results (time, travel expense and payroll);
    PCL3 - Applicant tracking data;
    PCL4 - Documents, Payroll year-end Tax data
    Database Tables PCLn
    · PCLn database tables are divided into subareas known as data clusters.
    · Data Clusters are identified by a two-character code. e.g RU for US payroll result, B2 for
      time evaluation result...
    · Each HR subarea has its own cluster.
    · Each subarea has its own key.
    Database Table PCL1
    · The database table PCL1 contains the following data areas:
      B1 time events/PDC
      G1 group incentive wages
      L1 individual incentive wages
      PC personal calendar
      TE travel expenses/payroll results
      TS travel expenses/master data
      TX infotype texts
      ZI PDC interface -> cost account
    Database Table PCL2
    · The database table PCL2 contains the following data areas:
      B2 time accounting results
      CD cluster directory of the CD manager
      PS generated schemas
      PT texts for generated schemas
      RX payroll accounting results/international
      Rn payroll accounting results/country-specific ( n = HR country indicator )
      ZL personal work schedule
    Database Table PCL3
    · The database table PCL3 contains the following data areas:
      AP action log / time schedule
      TY texts for applicant data infotypes
    Data Management of PCLn
    · The ABAP commands IMPORT and EXPORT are used for management of read/write to
      database tables PCLn.
    · A unique key has to be used when reading data from or writing data to the PCLn.
      Field Name KEY Length Text
      MANDT X 3 Client
      RELID X 2 Relation ID (RU,B2..)
      SRTFD X 40 Work Area Key
      SRTF2 X 4 Sort key for dup. key
    Cluster Definition
    · The data definition of a work area for PCLn is specified in separate programs which comply  
       with fixed naming conventions.
    · They are defined as INCLUDE programs (RPCnxxy0). The following naming convention applies:
       n = 1 or 2 (PCL1 or PCL2)
       xx = Relation ID (e.g. RX)
       y = 0 for international clusters or country indicator (T500L) for different country cluster
    Exporting Data (I)
    · The EXPORT command causes one or more 'xy' KEY data objects to be written to cluster xy.
    · The cluster definition is integrated with the INCLUDE statement.
    REPORT ZHREXPRT.
    TABLES: PCLn.
    INCLUDE: RPCnxxy0. "Cluster definition
    Fill cluster KEY
    xy-key-field = .
    Fill data object
    Export record
    EXPORT TABLE1 TO DATABASE PCLn(xy) ID xy-KEY.
       IF SY-SUBRC EQ 0.
           WRITE: / 'Update successful'.
       ENDIF.
    Exporting Data (II)
    . Export data using macro RP-EXP-Cn-xy.
    · When data records are exported using macro, they are not written to the database but to a  
      main memory buffer.
    · To save data, use the PREPARE_UPDATE routine with the USING parameter 'V'.
    REPORT ZHREXPRT.
    *Buffer definition
    INCLUDE RPPPXD00. INCLUDE RPPPXM00. "Buffer management
    DATA: BEGIN OF COMMON PART 'BUFFER'.
    INCLUDE RPPPXD10.
    DATA: END OF COMMON PART 'BUFFER'.
    RP-EXP-Cn-xy.
    IF SY-SUBRC EQ 0.
        PERFORM PREPARE_UPDATE USING 'V'..
    ENDIF.
    Importing Data (I)
    · The IMPORT command causes data objects with the specified key values to be read from
       PCLn.
    · If the import is successful, SY-SUBRC is 0; if not, it is 4.
    REPORT RPIMPORT.
    TABLES: PCLn.
    INCLUDE RPCnxxy0. "Cluster definition
    Fill cluster Key
    Import record
    IMPORT TABLE1 FROM DATABASE PCLn(xy) ID xy-KEY.
       IF SY-SUBRC EQ 0.
    Display data object
       ENDIF.
    Importing data (II)
    · Import data using macro RP-IMP-Cn-xy.
    · Check return code SY-SUBRC. If 0, it is successful. If 4, error.
    · Need include buffer management routines RPPPXM00
    REPORT RPIMPORT.
    *Buffer definition
    INCLUDE RPPPXD00.
    DATA: BEGIN OF COMMON PART 'BUFFER'.
    INCLUDE RPPPXD10.
    DATA: END OF COMMON PART 'BUFFER'.
    *import data to buffer
    RP-IMP-Cn-xy.
    *Buffer management routines
    INCLUDE RPPPXM00.
    Cluster Authorization
    · Simple EXPORT/IMPORT statement does not check for cluster authorization.
    · Use EXPORT/IMPORT via buffer, the buffer management routines check for cluster
      authorization.
    Payroll Results (I)
    · Payroll results are stored in cluster Rn of PCL2 as field string and internal tables.
      n - country identifier.
    · Standard reports read the results from cluster Rn. Report RPCLSTRn lists all payroll results;
      report RPCEDTn0 lists the results on a payroll form.
    Payroll Results (II)
    · The cluster definition of payroll results is stored in two INLCUDE reports:
      include: rpc2rx09. "Definition Cluster Ru (I)
      include: rpc2ruu0. "Definition Cluster Ru (II)
    The first INCLUDE defines the country-independent part; The second INCLUDE defines the country-specific part (US).
    · The cluster key is stored in the field string RX-KEY.
    Payroll Results (III)
    · All the field string and internal tables stored in PCL2 are defined in the ABAP/4 dictionary. This
      allows you to use the same structures in different definitions and nonetheless maintain data
      consistency.
    · The structures for cluster definition comply with the name convention PCnnn. Unfortunately, 
       'nnn' can be any set of alphanumeric characters.
    *Key definition
    DATA: BEGIN OF RX-KEY.
         INCLUDE STRUCTURE PC200.
    DATA: END OF RX-KEY.
    *Payroll directory
    DATA: BEGIN OF RGDIR OCCURS 100.
         INCLUDE STRUCTURE PC261.
    DATA: END OF RGDIR.
    Payroll Cluster Directory
    · To read payroll results, you need two keys: pernr and seqno
    . You can get SEQNO by importing the cluster directory (CD) first.
    REPORT ZHRIMPRT.
    TABLES: PERNR, PCL1, PCL2.
    INLCUDE: rpc2cd09. "definition cluster CD
    PARAMETERS: PERSON LIKE PERNR-PERNR.
    RP-INIT-BUFFER.
    *Import cluster Directory
       CD-KEY-PERNR = PERNR-PERNR.
    RP-IMP-C2-CU.
       CHECK SY-SUBRC = 0.
    LOOP AT RGDIR.
       RX-KEY-PERNR = PERSON.
       UNPACK RGDIR-SEQNR TO RX-KEY-SEQNO.
       *Import data from PCL2
       RP-IMP-C2-RU.
       INLCUDE: RPPPXM00. "PCL1/PCL2 BUFFER HANDLING
    Function Module (I)
      CD_EVALUATION_PERIODS
    · After importing the payroll directory, which record to read is up to the programmer.
    · Each payroll result has a status.
      'P' - previous result
      'A' - current (actual) result
      'O' - old result
    · Function module CD_EVALUATION_PERIODS will restore the payroll result status for a period
       when that payroll is initially run. It also will select all the relevant periods to be evaluated.
    Function Module (II)
    CD_EVALUATION_PERIODS
    call function 'CD_EVALUATION_PERIODS'
         exporting
              bonus_date = ref_periods-bondt
              inper_modif = pn-permo
              inper = ref_periods-inper
              pay_type = ref_periods-payty
              pay_ident = ref_periods-payid
         tables
              rgdir = rgdir
              evpdir = evp
              iabkrs = pnpabkrs
         exceptions
              no_record_found = 1.
    Authorization Check
       Authorization for Persons
    ·  In the authorization check for persons, the system determines whether the user has the 
       authorizations required for the organizational features of the employees selected with
       GET PERNR.
    ·  Employees for which the user has no authorization are skipped and appear in a list at the end
       of the report.
    ·  Authorization object: 'HR: Master data'
    Authorization for Data
    · In the authorization check for data, the system determines whether the user is authorized to
      read the infotypes specified in the report.
    · If the authorization for a particular infotype is missing, the evaluation is terminated and an error
      message is displayed.
    Deactivating the Authorization Check
    · In certain reports, it may be useful to deactivate the authorization check in order to improve
      performance. (e.g. when running payroll)
    · You can store this information in the object 'HR: Reporting'.
    these are the main areas they ask q?

  • IDoc not getting processed in Real Time job

    Hello,
    The configurations are done as per the below link to receive iDoc from SAP in BODS.
    How to Receive IDOC in BODS from SAP (Outbound to SAP)
    An iDoc is sent and Real Time Services on management console shows that request is received successfully. But I cannot see output xml file created. When I checked the real time services and client interfaces, I could see all are in green.
    But the real time job status shows its in running state. It has been in running state for 2 days. Trace shows as below. There is no error message.
    14.2) 04-24-14 18:47:41 (18417:1140171104)      JOB: Optimizing job <JOB_BODS_RT_IDocMatmas01>.
    (14.2) 04-24-14 18:47:47 (18417:1140171104)      JOB: Job <JOB_BODS_RT_IDocMatmas01> is started.
    (14.2) 04-24-14 18:47:47 (18417:1140171104) DATAFLOW: Data flow <DF_RT_IDocMatmas01> is initialized.
    (14.2) 04-24-14 18:47:47 (18417:1140171104) DATAFLOW: Data flow <DF_RT_IDocMatmas01> using IN MEMORY Cache.
    (14.2) 04-24-14 18:47:48 (18417:1140171104) DATAFLOW: Data flow <ActaDefaultNotificationSender> is initialized.
    (14.2) 04-24-14 18:47:49 (18417:1140171104) DATAFLOW: Data flow <ActaDefaultNotificationSender> using IN MEMORY Cache.
    Can anyone please let me know if there is anything missing in the configuration? How do I resolve this issue?
    DS version: 4.2

    Hi,
    Try to reimport the metadata of IDOC into PI in idx2. Also check the port definition in ECC and whether the segment is released or not. Please go through the below discussions it may help you.
    EDISDEF: Port XXX segment defn E2EDK35 in IDoc type ORDERS05 CIM type
    IDoc 0000000000181226 was saved but cannot or should not be sent
    Regards,
    Priyanka

Maybe you are looking for