Calculation logic for date with respect to time.

Dear all,
In BI, i want to get the time for two different dates - how to caculate this
For example
Initial date :- 01.08.2007
Final date :- 24.08.2007  - I want system to calculate the total hours between these dates - the logic should also be applicable if years,months and days are different.
Regards,
M.M

Dear ....
By subtracting final from the initial we will be able to get if the difference is only in days but if the month or year is differing will the same logic hold good - kindly explain?
Regards,
M.M

Similar Messages

  • My waveform chart is suppose to be with respect to time but I've noticed it is plotted with respect to the number of times through my VI

    I was under the assumption that waveform charts in LabVIEW 6i atomaticly displayed wired data with respect to time in seconds. However, I noticed that my chart plots data with respect to the number of data points my VI has generated. Is there anyway to fix this so that the chart actually plots vs. time. I've attached VI. The CIN is to a Computerboards Inc, thermocouple. Any help would be appreciated.
    Attachments:
    tepTInScan.VI ‏73 KB

    It appears that the chart is set to display relative time but since you don't have the acquisition in any kind of a loop, your relative time is pretty meaningless. You've got the dX of the chart set to be one, so yeah, at the beginning, the x scale numbers are going to be the same as the number of aquisitions until you've run it more than 60 times. At this point, you'll see the x scale show minutes and seconds. If you want absolute time for the x scale, look at the examples Real-Time Chart.vi or Variable-Time Chart XY.vi.

  • Validation of service entry date with respect to po date

    i want to validate the service entrysheet date with respect to Purchase order date.
    service entry sheet should not be before po date(doc date).
    SRVESSR- name of user exit for service entry sheet.
    but it is not triggering ,
    suggest me how to do .
    if sy-tcode = 'ML81N'.
      if i_ekpo-werks = '5000' or i_ekpo-werks = '4000' or i_ekpo-werks = '6000'.
        if C_ESSR-BUDAT  < I_EKKO-BEDAT.
          message 'Service Entry sheet Psting data should not be before PO Date' type 'E'.
        endif.
      endif.

    Hi shaiz.
    Try to under the concept of schduling During MRP run,
    Scheduling will be done  for FG with respect to  scheduling data maintained for FG material. It will not consider the planned delevery time of its child  materail.
    Requirement date is created for FG only, on this date FG should be deliverd to stock.
    u have to plan  raw material to meet requirement date.
    System will not  reschedule FG order automatically if raw delivery time is delayed it will give exception message only to reschedule. You have to reschedule FG order with respect to delivery date of ROH  manually by doing material avilablity check.
    I hope this should have  given some idea
    Regards
    Pradeep

  • Metagen is skipped sometimes complaining because all output files are up-to-date with respect to the input files

    Hi,
    In one of my dll (VS2010), the process Skipping target "MetaGen" because all output files are up-to-date with respect to the input files.
    it is at  random.  I have checked the time stamps of the dll and metagen.write.1.tlog both having the time stamp on the failed build.  But all files in release folder did not have the same time.  They are one hour earlier than the dll
    and the  metagen.write.1.tlog.
    On the good build  it said the input file is is newer than output file ".\Release\\metagen.write.1.tlog" and allfiles in the release folder have same time stamp.
    Any clue to solve this ?
    Regards,

    Hi,
    Based on your post, I am not very clear about your issue. I feel your issue is out of support range of VS General Question forum which mainly discusses
    the usage of Visual Studio IDE such as WPF & SL designer, Visual Studio Guidance Automation Toolkit, Developer Documentation and Help System
    and Visual Studio Editor.
     Could you please provide us more information so that we resolve your issue better?
    What type of project are you working with? C++ MFC project?
    How was the
    metagen.write.1.tlog file generated in your project?
    What error messages did you get?
    What is the 'input file'?
    If possible, please provide us a sample project to reproduce your issue.
    Thanks,
    We are trying to better understand customer views on social support experience, so your participation in this interview project would be greatly appreciated if you have time. Thanks for helping make community forums a great place.
    Click
    HERE to participate the survey.

  • Looking for help with respect to configuring MS Exchange server to handle attachments over 10 MB for forwarding to Salesforce (Email-to-case).

    Looking for help with respect to configuring MS Exchange server to handle attachments over 10 MB for forwarding to Salesforce (Email-to-case).
    Problem - SFDC does not create cases from emails that have more than 10 MB of attachments. Our clients will not go-live if their clients cannot send in emails with attachments over 10 MBs
    Potential resolution - Configure MS exchange to strip off the attachments(if over 10 MB) and store it in a public folder, forward the email to Salesforce (so the case gets created or the email
    is associated to an existing case), the client should have some way to know if the attachments were stripped off and should be able to dlownload the attachments and continue with case resolution.
    Any help is appreicated!
    Thanks

    Hi,
    From your description, you want to achieve the following goal:
    Configure Exchange to filter the attachments if the size is over 10 MB and store it in a public folder, and then forward the email to Salesforce.
    Based on my knowledge, I'm afraid that it can't be achieved. Exchange can filter messages with attachments, but it couldn't store these attachments on public folder automatically. Also, I don't see any transport rule can do it.
    Hope my clarification is helpful.
    Best regards,
    Amy Wang
    TechNet Community Support

  • Post Calculation Computation for Date/Timestamps

    Hello,
    I would like to know how to apply Post Calculation Computation for Dates/Timestamps?
    I had trouble trying to add this to the Post Calculation Computation area:
    TO_CHAR(:P1_TARGETDATE, :DEFAULT_DATE_TIME_MASK);
    or should it be something more like:
    SELECT
    TO_CHAR(:P1_TARGETDATE, :DEFAULT_DATE_TIME_MASK) FROM DUAL;
    What would be the correct code to allow me to format a date/Timestamp where a Application Variable of DEFAULT_DATE_TIME_MASK could be used as well. (I dont want to use the existing Format Mask.)

    Just a hunch, but you might be having a problem because of the double quotes. Oracle likes single quotes. You also might want to forgo the quotes entirely.
    As for an invalid number, is it possible that one of the values that you are adding might be NULL. You can't add NULL to a number or you get an error. There are two solutions to that. One would be to do an NVL around each number you are adding (changing it to zero if the value is NULL). The second option (and easier option) is to use the SUM function. Mathematical functions like SUM and AVG ignore those values that are NULL.

  • Moving average price calculation logic of material with Price Control "S"

    Dear Gurus,
    As you know that there is a Moving price and standard price icon in the material master.
    I want to understand the calculation logic of the moving average price of the materials having price control "S"
    How the system calculates the MAP for standard price materials? The receipt from the process order I suppose is valuated at the process order cost after settlement, but if the issue has hapened for the material does the system recalculate the issue price also?
    Below is the sample process order receipt and issue scenario:
    Receipt from process order 161000000223
    300 kg and GR at standard cost value is Rs 5892
    Issue to process order 162000000294
    250 kg and GI at standard cost value is Rs 4910.
    Thus the balance at period end is 50 kg and balance at standard cost value is Rs 982.
    Here in process order 161000000223 the actual cost is 10 Rs. Then how will the system calculate the MAP?
    Thanking You,
    Amit Dhanurdhari

    Hi
    Try the following calculations,
    One of them will work depending upon your version and support pack
    Expected MAP calculation ( As with price control V)
    = ((Qty before transaction*MAP + Transaction Qty * transaction
    Price)) / quantity after transaction
    New Method ( when price control is S)
    = Old MAP + (Price Variance w.r.t Standard Price/Qty after transaction)
    Also go through the following notes,
    Note 1225167
    1253944
    518485 FAQ: Valuation of goods movements
    212286 Overview note: Valuation during goods movements
    209864 Moving average price is disproportionately large
    202166 Collective note: Statistical moving average price
    185961 Moving Average Price Calculation
    I have done extensive research on this, let me know if you need to know something specific.

  • Csv file reading and voltage and current plotting with respect to time samples XY plotting

    Hallo,
             I've been struggling with reading  a comma separated value (csv) file from another instrument (attached). I need to plot this data for analysis. I have 5 column of data with numbers of rows, the first three row is the information of the measurement. I want to read 4th row as string and rest of the row as numbers. I want to plot 2nd column (i1)  with respect to TIMESTAMP; 4th column(u2) wrt TIMESTAMP. And finally plotting i1 (x-axis) vs.. u2 (y-axis) in labview. Could anyone help me.
    In excel its so easy to plot  but I don't know how its done in labview.
    Attachments:
    labview forum test.csv ‏30 KB
    excel plot.jpg ‏88 KB

    Start by opening the file.  Then use the Read Text File function.  You can right-click on it and configure it to read lines.  First make it read 3 lines (this is your extra header data).  Then make it read a single line.  That will give you your channel names.  Then read the rest of the file (disable the read by line and wire a -1 into the number of bytes to read).  Then use the Spreadsheet String to Array function to give you your data.
    I would recommend going through the LabVIEW tutorials if you are really new at this.
    LabVIEW Basics
    LabVIEW 101
    There are only two ways to tell somebody thanks: Kudos and Marked Solutions
    Unofficial Forum Rules and Guidelines

  • Optimal read write performance for data with duplicate keys

    Hi,
    I am constructing a database that will store data with duplicate keys.
    For each key (a String) there will be multiple data objects, there is no upper limit to the number of data objects, but let's say there could be a million.
    Data objects have a time-stamp (Long) field and a message (String) field.
    At the moment I write these data objects into the database in chronological order, as i receive them, for any given key.
    When I retrieve data for a key, and iterate across the duplicates for any given primary key using a cursor they are fetched in ascending chronological order.
    What I would like to do is start fetching these records in reverse order, say just the last 10 records that were written to the database for a given key, and was wondering if anyone had some suggestions on the optimal way to do this.
    I have considered writing data out in the order that i want to retrieve it, by supplying the database with a custom duplicate comparator. If I were to do this then the query above would return the latest data first, and I would be able to iterate over the most recent inserts quickly. but Is there a performance penalty paid on writing to the database if I do this?
    I have also considered using the time-stamp field as the unique primary key for the primary database instead of the String, and creating a secondary database for the String, this would allow me to index into the data using a cursor join, but I'm not certain it would be any more performant, at least not on writing to the database, since it would result in a very flat b-tree.
    Is there a fundamental choice that I will have to make between write versus read performance? Any suggestions on tackling this much appreciated.
    Many Thanks,
    Joel

    Hi Joel,
    Using a duplicate comparator will slow down Btree access (writes and reads) to
    some degree because the comparator is called a lot during searching. But
    whether this is a problem depends on whether your app is CPU bound and how much
    CPU time your comparator uses. If you can avoid de-serializing the object in
    the comparator, that will help. For example, if you keep the timestamp at the
    beginning of the data and only read the one long timestamp field in your
    comparator, that should be pretty fast.
    Another approach is to store the negation of the timestamp so that records
    are sorted naturally in reverse timestamp order.
    Another approach is to read backwards using a cursor. This takes a couple
    steps:
    1) Find the last duplicate for the primary key you're interested in:
      cursor.getSearchKey(keyOfInterest, ...)
      status = cursor.getNextNoDup(...)
      if (status == SUCCESS) {
          // Found the next primary key, now back up one record.
          status = cursor.getPrev(...)
      } else {
          // This is the last primary key, find the last record.
          status = cursor.getLast(...)
      }2) Scan backwards over the duplicates:
      while (status == SUCCESS) {
          // Process one record
          // Move backwards
          status = cursor.getPrev(...)
      }Finally another approach is to use a two-part primary key: {string,timestamp}.
    Duplicates are not configured because every key is unique. I mention this
    because using duplicates in JE has more overhead than using a unique primary
    key. You can combine this with either of the above approaches -- using a
    comparator, negating the timestamp, or scanning backwards.
    --mark                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                               

  • BAPI for MIGO with respect to Outbound delivery or Stk trfer pur order

    please let me know is there any standard BAPI for MIGO transaction with respect to Outbound delivery or Stock transfer purchase order

    try these...
    BAPI_OUTB_DELIVERY_CONFIRM_DEC
    BAPI_OUTB_DELIVERY_SAVEREPLICA
    BAPI_OUTB_DELIVERY_SPLIT_DEC

  • Avoiding change pointers for data with status planned

    Hi,
    we have to distribute HCM-data, standard infotypes and customer infotypes, into other SAP sytems. One of the receiving systems must not get data with the status planned (istat = 2).
    Which would be the best way to avoid planned data in the receiving system? Is it possible to avoid creating change pointers for planned data in the delivering system or have I to delete the planned data in the receiving system while processing the idocs via user-exit?
    Best regards
    Stefan

    Dear Stefan,
    You can create the change pointers manually from the IDOC data (e.g.with  FM CHANGE_POINTERS_CREATE) during the inbound processing within of the customer-exits of enhancement RHALE001 or within an own BAdI implementation of BAdI HRALE00INBOUND_IDOC.
    Or else you can use method IDOC_DATA_FOR_RECEIVER_MODIFY of BAdI HRALE00OUTBOUND_IDOC. In this case please make sure that note 1292241 is implemented.
    Hope it helps,
    Christine

  • How to access date with respective to different language

    Hello,
    In my database i stored the date in the following format1992-03-28 13:07:53.000.
    In My application user can use upto 8 languages.Now the problem is when the user logged in with english the date format will be like 5/26/08(dd/mm/yy).but me convert this format to 05/26/2008(dd/mm/yyyy) through and use the following syntax for constrains dateformat(MaoDate,'mm/dd/yyyy') = '05/26/2008'
    Now my problem is for some other language the date format is like Arr 20 07 or 25.05.92.But me wrote the code conversion only for the format mm/dd/yy.
    Is there any general way to acces the date for all the languages?
    Please help Me.
    Thank You!!!

    Hello sathish,
    Now the problem is i will get the string like 5/8/08 or Jan 21 07 .... from My GUI. But me wrote code only to convert from 5/8/08 to 05/08/2008...Bcz date comparison was happening in dd/mm/yyyy format.
    But whenever i am getting data like Jan 21 07 from GUI my conversion is not happening.And now the date constrain is dateformat(MaoDate,'mm/dd/yyyy') = 'Jan 21 07' i mean my code checking the date format with dd/mm/yyyy with date Jan 21 07.
    So i need to write a code to convert all the regional format to dd/mm/yyyy.So is there any common way to comapre the date with all regional format!!!!!!
    Thank You!!

  • Logic For Data Fetching

    I am doing a dump program for qp02 change inspection plan and  unable to fetch the data for qualitative and quantitative fields these fields can be x or '' . iam not getting the fields in plmk table..when i changed these fields STEUERKZ in plmk table gets reflected by adding
    X= . i think there is a different logic for it.If any one
    knows then pls reply me.
    regards.
    Sanchit Sharma

    Hi KR,
    if you want particular items from the table VBREVE (Revenue Recognition: Revenue Recognition Lines)
    Pass all the keyfields in where condition..
    try to provide maximum field values from below list in where condition... if you are not be able to get it directly then try to get it from some other tables and then paas it here...
    VBELN ==> Sales Document
    POSNR ==> Sales Document Item
    SAKRV ==> G/L Account Number
    BDJPOPER ==> Posting year and posting period (YYYYMMM format)
    POPUPO ==> Period sub-item
    VBELN_N ==> Subsequent sales and distribution document
    POSNR_N ==> Subsequent item of an SD document
    also have a look on other tables related to this...
    VBREF
    VBREVAC
    VBREVC
    VBREVE
    VBREVK
    VBREVR
    Hope it will solve your problem..
    Thanks & Regards
    ilesh 24x7
    ilesh Nandaniya

  • Vlan for voice and vlan for data with diferent ISP best choice of config??

    Hello everyone,
    Im, Oscar
    At our company we have a redundant ISP connection to two separate ISP's.
    We are also using VoIP on the network.
    Currently one ISP connection is used primarily and the second on is just used as a backup.
    I was wondering if it is possible to use the first connection primarily for data traffic and the second connection for VoIP traffic?
    We use different VLAN's for voice and data.
    Any help would be appreciated.

    Disclaimer
    The Author of this posting offers the information contained within this posting without consideration and with the reader's understanding that there's no implied or expressed suitability or fitness for any purpose. Information provided is for informational purposes only and should not be construed as rendering professional advice of any kind. Usage of this posting's information is solely at reader's own risk.
    Liability Disclaimer
    In no event shall Author be liable for any damages whatsoever (including, without limitation, damages for loss of use, data or profit) arising out of the use or inability to use the posting's information even if Author has been advised of the possibility of such damage.
    Posting
    Yes, for egress.  Ingress is "it depends".
    You could also consider using both links for both kinds of traffic.

  • Write custom logic for data archiving using Open text(SAP archiving)

    Hi,
    We are currently doing archiving for sales documents and billing documents using Open text. As part of this we need to have a search functionality based on Sales document,customer number, customer address etc...
    since customer address doesn't exist in sales order tables we need to write logic so that the document gets archived and retrieved using the address.
    can you please help me out in how and where to write logic for this...can this be achieved?
    i observed an area on Attribute Object for a Node which gives an option for Event and User exit. can this be used..if so how?
    Thanks,
    Anil

    hi,
    please write more detailed information what do you really need...

Maybe you are looking for