Finding the record with the latest date

Post Author: sconlon
CA Forum: General
Hi,I am using CR 8 with an Access database which has a table that holds (multiple) sales records for each of our clients. I want to create a report that lists only the latest sale from each client.  I am thinking that I first need to group by client and then create a sub-report that prints the necessary details of sales record with the latest date of sale.  It's this last bit that I cannot figure out how to do.  Can anyone suggest a way of doing this?Many thanks,sconlon

Post Author: Jagan
CA Forum: General
Charliy's suggestion if fine, and does require sorting and displaying in the group header/footer according to the sort order. In case it's not clear, when you print in the group header you are using the first record's data; when you print in the group footer you are using the last record's data (first and last within the group that is).
Personally, I'd group on the client have a group selection formula of:
{table.date_field} = maximum({table.date_field}, {table.client_id})
so that there's no date sort required as you'll only get one record per client group - assuming that each record has a unique date of course. I'd probably still suppress the details section and display in the group footer, but it wouldn't matter if you use the details section either.
I believe the benefit with the group selection formula is that any summary formulas, running totals etc. you use would only have the one record per client to work on, rather than including all of the suppressed records too.

Similar Messages

  • DTP error: No Message: Filter Out New Records with the Same Key

    Hello SDN,
    I have looked up solutions for this problem, but didnt find anything suitable. I am getting this error in multiple places: when i load pure master data into IOs like 0VENDOR, 0PLANT etc. or when I load DSOs like 0FIAP_O03.
    When I checked the load in sdn, one of the solutions was to apply note: 1489178 , but in this case, the user had nothing selected in his 'semantic key' list. Whereas, in my case.. 0PLANT is checked ...eventhough the other fields are disabled. So, I think my DTP is accurate.. as 'plant' is the only key field here.  But it still fails .. and so does 0VENDOR load. Moreover... the DSO load is crucial for our reports in Prod and its failure is a bigger issue for us.
    Has anyone seen this problem? Is there another solution other than the above note?
    Regards.

    Hello Arvind,
    If I select that option, will it delete one record if it finds 2 records with the same semantic key in 1 data packet. Also, will it delete a record if it finds that there is already a record existing in the data target with the same semantic key.
    I think, both of the above scenarios will result in incorrect data. From my understanding, 'Handle Duplicate Keys' is mostly used to filter out master data.. as in some instances.. you might get duplicate entries.
    But, in my case, the same error is happening in the case of data load from 0fi_ap_4 --> 0fiap_o03. What happens is... the overall status of the DTP becomes green.. but when you look at details section, the process stops at this message which is in yellow. Because of this, the DTP status in the PC remains yellow.. and finally turns red.. and the PC turns red.
    From my knowledge, it appears more like a bug. I was hoping to find a SAP NOTE which has some manual instructions that can resolve the issue, since the above NOTE asks for a support pack upgrade.. which would need a whole lotta testing.
    Regards.

  • How to filter the records in the process of loading data into ODS

    Hi All,
    I am doing some data extraction from R/3 system. But I need to ignore(don't need to load) the records with a date field (Purchase Order date) that is blank.
    So how can I filter out the records with the blank date field?
    Some one suggested me to write some select statement (looping the data packages) in the start routine but I am not that experianced with the ABAP routines. So can anyone provide some sample format of the routine or can you suggest me some other way to  filter out this data.

    Ram Kumar,
    I used this in the start routine but it gives me an error saying ...
    No component exists with the name "TRAN_STRUCTURE-BEDAT"
    <b>delete datapak where TRAN_STRUCTURE-BEDAT = '' OR
                         TRAN_STRUCTURE-BADAT = '' OR
                         TRAN_STRUCTURE-ERDAT = '' OR
                         TRAN_STRUCTURE-FRGDT = '' OR
                         TRAN_STRUCTURE-UDATE = ''.</b>
    I have taken out the TRAN_STRUCTURE part and just typed
    delete datapak where BEDAT = '' OR
                         BADAT = '' OR
                         ERDAT = '' OR
                         FRGDT = '' OR
                         UDATE = ''.
    This above statement has no syntax errors. but this statement doesn't filter out the records with a the date fields that are blank.
    Please help me out here.
    This is kind of urgent.
    Thanks,

  • Send emails from a csv file (grouping records with the same email addresses)

    Writing a script to send emails from a csv which contains record details and email addresses using the Send-Mailmessage cmdlet.
    That part is no trouble, however, the csv file contain records that can have the same email address.
    Is there a way/method I could send the records with the same email address once instead of sending individual email messages?
    Thanks
    Data in the csv file
    Record number, description, email
    1234, Test 1, [email protected]
    5678, Test 2, [email protected]
    1245, Test 3, [email protected]
    4578, Test 4, [email protected]
    $data = Import-csv c:\records.csv
    ForEach($address in $data)
    Send-mailmessage -To $address.emails -from [email protected] -subject "List of records" -Body $data -SmtpServer 192.168.1.1

    The following code does what you are looking for:
    $data = Import-csv c:\records.csv$mx = "192.168.1.1"
    $subject = "List of records"
    $from = "[email protected]"
    $data| Group-Object email|Select-Object Name, @{n='msg'; e={$_.Group| Select-Object -Property "record number", description|ConvertTo-Csv -notypeinformation|Out-String}}|
    ForEach-Object {Send-MailMessage -to $_.Name -Body $_.msg -SmtpServer $mx -Subject $subject -From $from}
    I also recommend no forcing everything into one line.  "One-liner" means "one pipeline" not putting all code on one unreadable line.
    Your code should look like this:
    $data|
    Group-Object email|
    Select-Object Name, @{
    n='msg'; e={
    $_.Group| Select-Object -Property "record number", description|
    ConvertTo-Csv -notypeinformation|Out-String
    }|
    ForEach-Object {
    Send-MailMessage -to $_.Name -Body $_.msg -SmtpServer $mx -Subject $subject -From $from
    Now we can see the code and see that you are still unnecessarily converting back and forth.
    To get the group look at how I did it.
    $body=$_.Group | Format-Table |Out-String
    Isn't that much easier?
    Once you master the pipeline these things will become second nature.  In PowerShell it is not necessary to write lots of code most of the time.
    ¯\_(ツ)_/¯

  • 2lis_02_scl extract records with wrong posting date for Good Receipts.

    Hi Experts,
    We are currently having issue of mismatch between BW Schedule Line data with R/3 values for the Goods Receipts posting date updating incorrectly in to BW.
    Example.
    In table EKBE purchase order history we have following records.
    MANDT EBELN        EBELP   ZEKKN VGABE GJAHR BELNR      BUZEI BEWTP BWART BUDAT      MENGE
    501   5600453404   00010   00    1     2010  5012473031 0001  E     101   23.01.2010        1.250,000
    501   5600453404   00010   00    1     2010  5012473031 0002  E     101   23.01.2010        1.250,000
    501   5600453404   00010   00    1     2010  5012473031 0003  E     101   23.01.2010        1.250,000
    501   5600453404   00010   00    1     2010  5012693310 0001  E     101   26.02.2010        1.250,000
    Which means we have on posting date of 23.01.2010 1250*3 i.e. 3750 quantity of Goods Receipts.
    However when we check the extractor we get multiple records in internal table C_T_DATA and in psa.
    Line BWV ETENR   SLFDT              MENGE    ROCA BEDAT                            BUDAT          EBELN
           ORG                                                        NCEL
    1     001     0001     20100125     3750.000          20100113     ZNB     F     00000000     5600453404
    2     001     0002     20100226     1250.000          20100113     ZNB     F     00000000     5600453404
    3     002     0001     20100125     3750.000     X     20100113     ZNB     F     20100123     5600453404
    4     002     0001     20100125     3750.000          20100113     ZNB     F     20100123     5600453404
    5     003     0001     20100125     3750.000     X     20100113     ZNB     F     20100127     5600453404
    6     003     0001     20100125     3750.000          20100113     ZNB     F     20100127     5600453404
    7     002     0001     20100125     3750.000     X     20100113     ZNB     F     20100226     5600453404
    8     002     0001     20100125     3750.000          20100113     ZNB     F     20100226     5600453404
    9     002     0002     20100226     1250.000     X     20100113     ZNB     F     20100226     5600453404
    As can be seen we have record no 8 for ETENR (Schedule line 1) with posting date 26.02.2010 and another record with posting date 23.01.2010.
    Since we are getting 2 records the record with incorrect posting date overwrites the record with correct one.
    Any idea if this could be a standard extractor problem or any other way to resolve this issue.
    Any help would be appreciated.

    First of all, are you using a staging DSO? (You should ideally)
    If yes, is it a Write Optimized DSO? (Again, this is ideal)
    If its a standard DSO, the values maybe over-writing upon activation.
    You have 3 records (quantity = 1250 * 3) that have been receipted on 23.01.2010, where Posting Date = 23.01.2010.
    You also have a record (quantity = 1250 * 1) that has been receipted on 26.01.2010. Posting date = 26.01.2010.
    Now, in RSA3 & in PSA you can see more records than intended.
    This is because you have before & after images. (ROCANCEL = X or blank).
    ROCANCEL = X --> Before Image (record before change)
    ROCANCEL = blank --> After Image (record after change)
    This is a standard property of the extractor.
    Now, we also have something known as BWVORG (Process Keys). Each process, i.e. creating a PO, Goods Receipt, Invoice etc.. have different Process Keys.
    Creation - 001
    Goods Rcpt. - 002
    Invoice - 003
    We can see that record 8, i.e. BWVORG = 002 (GR) has been modified on 26.01.2010.
    That is why there is before image and after image.
    Which one should be the correct posting date? 23.01.2010?
    Normally in a Write optimized DSO, you will have all the records (before and after images & others as well).
    I hope this helps.
    Please let me know if otherwise.

  • Filter records with month and date

    Hi,
    i have records as follows in the table.
    custno trn_dt amount
    1 01-jan-06 100
    1 10-jan-06 200 --->
    1 15-feb-06 300
    2 15-dec-06 111
    2 15-jan-06 222 --->
    2 01-fen-06 333
    If i specify the runtime parameter as JAN-06 the output of the report should be with 2 records indicated with arrows above.
    the query should pick up the record with the highest date for a specific customer.
    in the above example for customer "1" 10-jan-06 is highest date and for customer "2" 15-jan-06 is the highest date.
    hope the requirement is clear.
    thanks in advance

    SELECT custno, trn_dt, amount
    FROM TABLE
    WHERE (custno,trn_dt) IN
    (SELECT custno, MAX(trn_dt)
    FROM TABLE
    GROUP BY custno)
    There are many examples on this in SQL and PL/SQL forum.

  • How to display record with most recent date in sapui5?

    Hi
    I have a local json data with me, in which I have number of records.
    And in each record I have an "AENDATE" as a date property.
    Now I want to display the record with most recent date in the records.
    How I can I do it......????????
    Please help me with this.
    Thanks
    Sathish

    How about sorting your json model descending, and display only first (index:0) item?

  • Why is the recording on the cd not playable in itunes?

    I inserted a cd with a recording. the recording with the ending .WAV is not playable.
    can anyone help?
    I appreciate your response

    I get the same problem. I moved the backup files (B1) to another drive and then itunes generated new backup files (B2) for my iPhone. They seem to be of similar size. Due to my running out of space on my system drive, I've linked the backup folder to another drive. This is where I moved my original backups (B1). After I did that, the iPhone seems to sync but once again, the date of the backup did not change. I have no idea what the issue is. I'll save both B1 and B2.
    To be sure you've got an up to date backup, I recommend changing the name of the original backup folder to backup2 and then forcing itunes to make a new backup from scratch.
    On a windows machine, the backup folder is found at:
    C:\Users\(Username)\AppData\Roaming\Apple Computer\MobileSync\Backup

  • How to pull the job with the latest effective date in PSFT?

    Hello,
    I am running the PSFT 9.1.1 connector workforcefullsync.
    I ran into a problem when the user has multiple job records. Each time that the user job is updated, a new job record is created. So, when the data comes from PSFT, for a user, we have multiple jobs records with different effective date (For example <EFFDT IsChanged="Y">1996-10-21</EFFDT>). OIM is only interested in the latest EFFDT.
    Question 1: is it something that we can set on the PeopleSoft side, so that only the job with the highest EFFDT is sent to OIM?
    Question 1b: is this is not easy to accomplish from the PSFT side, can I use transformation in the connector to only pull the job record with the highest effective date?
    Question 2: if the EFFDT shows a date in the future (HR wants to disable a person in the future), does OIM ignore this change until the sysdate is after that EFFDT date?
    Thanks
    PS: workforcefullsync worked fine if the file contains only 1 job per employee. When the xml files contain multiple job records, I got the error:
    ERROR QuartzWorkerThread-1 OIMCP.PSFTER - oracle.iam.connectors.psft.common.handler.impl.PSFTWorkForceSyncReconMessageHandlerImpl : handleMessage
    ERROR QuartzWorkerThread-1 OIMCP.PSFTER - 1
    ERROR QuartzWorkerThread-1 OIMCP.PSFTER - Description : 1
    ERROR QuartzWorkerThread-1 OIMCP.PSFTER - java.lang.ArrayIndexOutOfBoundsException: 1
    Edited by: user12049102 on Mar 22, 2010 12:06 PM

    Hello,
    When the PSFT team sends a message over to OIM to disable a user with today's date, the user got disabled fine in OIM. A reconciliation event is created.
    When the PSFT team sends a message over to OIM to disable a user with a future date for EFFDT, let's say 3/31/2010, no reconciliation event was created.
    Does OIM store this information somewhere so that it will process it later on?
    It's a good thing that OIM does not disable the user who has an EFFDT in the future, but we don't want that record to be forgotten.
    Please help.
    Thanks

  • One-to-many selfjoin removing records with the same ranking or with a substitute

    Sorry for my bad choice of discussion title, feel free to suggest me a more pertinent one
    I've rewritten post for clarity and following the FAQ.
    DB Version
    I'm using Oracle Enterprise 10g 10.2.0.1.0 64bit
    Tables involved
    CREATE TABLE wrhwr (
    wr_id INTEGER PRIMARY KEY,
    eq_id VARCHAR2(50) NULL,
    date_completed DATE NULL,
    status VARCHAR2(20) NOT NULL,
    pmp_id VARCHAR2(20) NOT NULL,
    description VARCHAR2(20) NULL);
    Sample data
    INSERT into wrhwr  VALUES (1,'MI-EXT-0001',date'2013-07-16','Com','VER-EXC','Revisione')
    INSERT into wrhwr VALUES  (2,'MI-EXT-0001',date'2013-07-01','Com','VER-EXC','Verifica')
    INSERT into wrhwr  VALUES (3,'MI-EXT-0001',date'2013-06-15','Com','VER-EXC','Revisione')
    INSERT into wrhwr  VALUES (4,'MI-EXT-0001',date'2013-06-25','Com','VER-EXC','Verifica')
    INSERT into wrhwr  VALUES (5,'MI-EXT-0001',date'2013-04-14','Com','VER-EXC','Revisione')
    INSERT into wrhwr  VALUES (6,'MI-EXT-0001',date'2013-04-30','Com','VER-EXC','Verifica')
    INSERT into wrhwr  VALUES (7,'MI-EXT-0001',date'2013-03-14','Com','VER-EXC','Collaudo')
    Query used
    SELECT *
      FROM (SELECT eq_id,
                   date_completed,
                   RANK ()
                   OVER (PARTITION BY eq_id
                         ORDER BY date_completed DESC NULLS LAST)
                      rn
              FROM wrhwr
             WHERE     status != 'S'
                   AND pmp_id LIKE 'VER-EX%'
                   AND description LIKE '%Verifica%') table1,
           (SELECT eq_id,
                   date_completed,      
                   RANK ()
                   OVER (PARTITION BY eq_id
                         ORDER BY date_completed DESC NULLS LAST)
                      rn
              FROM wrhwr
             WHERE     status != 'S'
                   AND pmp_id LIKE 'VER-EX%'
                   AND description LIKE '%Revisione%') table2,
           (SELECT eq_id,
                   date_completed,           
                   RANK ()
                   OVER (PARTITION BY eq_id
                         ORDER BY date_completed DESC NULLS LAST)
                      rn
              FROM wrhwr
             WHERE     status != 'S'
                   AND pmp_id LIKE 'VER-EX%'
                   AND description LIKE '%Collaudo%') table3
    WHERE     table1.eq_id = table3.eq_id
           AND table2.eq_id = table3.eq_id
           AND table1.eq_id = table2.eq_id
    Purpose of the above query is to selfjoin wrhwr table 3 times in order to have for every row:
    eq_id;
    completition date of a work request of type Verifica for this eq_id (table1 alias);
    completition date of a wr of type Revisione (table2 alias) for this eq_id;
    completition date of a wr of type Collaudo (table3 alias) for this eq_id;
    A distinct eq_id:
    can have many work requests (wrhwr records) with different completition dates or without completition date (date_completed column NULL);
    in a date range can have all the types of wrhwr ('Verifica', 'Revisione', 'Collaudo') or some of them (ex. Verifica, Revisione but not Collaudo, Collaudo but not Verifica and Revisione, etc.);
    substrings in description shouldn't repeat;
    (eq_id,date_completed) aren't unique but (eq_id,date_completed,description,pmp_id) should be unique;
    Expected output
    Using sample data above I expect this output:
    eq_id,table1.date_completed,table2.date_completed,table3.date_completed
    MI-EXT-001,2013-07-01,2013-07-16,2013-03-14 <--- for this eq_id table3 doesn't have 3 rows but only 1. I want to repeat the most ranked value of table3 for every result row
    MI-EXT-001,2013-07-01,2013-06-15,2013-03-14 <-- I don't wanna this row because table1 and table2 have both 3 rows so the match must be in rank terms (1st,1st) (2nd,2nd) (3rd,3rd)
    MI-EXT-001,2013-06-25,2013-06-15,2013-03-14 <-- 2nd rank of table1 joins 2nd rank of table2
    MI-EXT-001,2013-04-30,2013-04-14,2013-03-14 <-- 1st rank table1, 1st rank table2, 1st rank table3
    In vector style syntax, expected tuple output must be:
    ix = i-th ranking of tableX
    (i1, i2, i3) IF EXISTS an i-th ranking row in every table
    ELSE
    (i1, b, b)
    where b is the first available lower ranking of table2 OR NULL if there isn't any row  of lower ranking.
    Any clues?
    With the query I'm unable to remove "spurius" rows.
    I'm thinking at a solution based on analytic functions like LAG() and LEAD(), using ROLLUP() or CUBE(), using nested query but I would find a solution elegant, easy, fast and easy to maintain.
    Thanks

    FrankKulash ha scritto:
    About duplicate dates: I was most interested in what you wanted when 2 (or more) rows with the same eq_id and row type (e.g. 'Collaudo') had exactly the same completed_date.
    In the new results, did you get the columns mixed up?  It looks like the row with eq_id='MI-EXT-0002' has 'Collaudo' in the desciption, but the date appears in the verifica column of the output, not the collaudo  column.
    Why don't you want 'MI-EXT-0001' in the results?  Is it realted to the non-unique date?
    For all optimization questions, see the forum FAQ:https://forums.oracle.com/message/9362003
    If you can explain what you need to do in the view (and post some sample data and output as examples) then someone might help you find a better way to do it.
    It looks like there's a lot of repetition in the code.  Whatever you're trying to do, I suspect there's a simpler, more efficient way to do it.
    About Duplicate dates: query must show ONLY one date_completed and ignore duplicated. Those records are "bad data". You can't have 2 collaudos with the same date completed.
    Collaudo stands for equipment check. A craftperson does an equipment check once a day and, with a mobile app, update the work request related to equipment and procedure of preventive maintenance, so is impossibile that complete more than one check (Collaudo) in a day, by design.
    In the new results, it's my fault: during digitation I've swapped columns
    With "I don't want 'MI-EXT-0001'" I mean: "I don't want to show AGAIN MI-EXT-0001. In the previous post was correct the output including MI-EXT-0001.
    Regarding optimisation...
    repetition of
    LAST_VALUE ( 
                            MIN (CASE WHEN r_type = THEN column_name END) IGNORE NULLS) 
                         OVER (PARTITION BY eq_id ORDER BY r_num)  AS alias_column_name
    is because I don't know another feasible way to have all columns needed of table wrhwr in main query, maintaining the correct order. So i get them in got_r_type and propagate them in all the subquery.
    In main query I join eq table (which contains all the information about a specific equipment) with "correct" dates and columns of wrhwr table.
    I filter eq table for the specific equipment standard (eq_std column).
    efm_eq_tablet table and where clause
    AND e.eq_id = e2.eq_id 
              AND e2.is_active = 'S'; 
    means: show only rows in eq table that have an equal row in efm_eq_tablet table AND are active (represented by 'S' value in is_active column).
    About the tables v2, r2 and c2
              (SELECT doc_data, doc_data_rinnovo, eq_id 
                 FROM efm_doc_rep edr 
                WHERE edr.csi_id = '1011503' AND edr.doc_validita_temp = 'LIM') v2, 
              (SELECT doc_data, doc_data_rinnovo, eq_id 
                 FROM efm_doc_rep edr 
                WHERE     eq_id = edr.eq_id 
                      AND edr.csi_id = '1011504' 
                      AND edr.doc_validita_temp = 'LIM') r2, 
              (SELECT doc_data, doc_data_rinnovo, eq_id 
                 FROM efm_doc_rep edr 
                WHERE     edr.csi_id IN ('1011505', '1011507') 
                      AND edr.doc_validita_temp = 'LIM' 
                      AND adempimento_ok = 'SI') c2, 
    Those tables contains "alternate" dates of completition to be used when there isn't any wrhwr row for an eq_id OR when all date_completed are NULL.
    NVL() and NVL2() functions are used in main query in order to impletement this.
    The CASE/WHEN blocks inside main query implements the behavior of selecting the correct date based of the above conditions.

  • Reducing time while Filtering Out New Records with the Same Key

    Hi Experts,
    I have an issue .. I am trying to load data into 0MATERIAL . While loading via DTP the step u201CFilter Out New Records with the Same Keyu201D   takes much time for processing . For each datapackage it is taking around 30 minutes to filter out new records with the same key .  While for other master data it hardly takes 5 minutes.
    Any pointers how to reduce the timeu2026
    Thanks in advance
    Sam

    Hello,
    No there is no need to do that change then.
    Can you tell me are you doing a full load daily to 0MATERIAL object.
    why dont you use delta.
    Also if the DTP is of type FULL, then are you deleting the previous PSA requests.
    Maybe you can ask the basis guys to check with a trace what is really happening on the DB side during this 30 mins and then maybe we can find the needed fix.
    Regards,
    Shashank

  • Record with the DIMID 6 contains non-fitting time characteristics

    hello guys,
    I made a test on our transactional infocube in the RSRV and I got the followings messages:
    *Record with the DIMID 6 contains non-fitting time characteristics
    *No master data found for characteristic of test described subsequently
    *No master data found for characteristic of test described subsequently
    this happens in many DIMID's. an the infocube was partitioned.
    I executed the correct error buttom but nothing.
    any idea how to fix it?
    thks
    EV

    Hello Emiliano,
    did you find a solution?
    Regards Murat

  • I updated the IPAD with the latest update and now when I turned the IPAD on, I get the itunes logo and then nothing else.  How can I fix this to let me use my IPAD?

    I updated the IPAD with the latest update and now when I turned the IPAD on, I get the itunes logo and then nothing else.  How can I fix this to let me use my IPAD?

    YOU ARE IN RECOVERY MODE
    1. Turn off iPad
    2. Turn on computer and launch iTunes (make sure you have the latest version of iTune)
    3. Plug USB cable into computer's USB port
    4. Hold Home button down and plug the other end of cable into docking port.
    DO NOT RELEASE BUTTON until you see picture of iTunes and plug
    5. Release Home button.
    ON COMPUTER
    6. iTunes has detected iPad in recovery mode. You must restore this iPad before it can be used with iTunes.
    7. Select "Restore iPad"...
    Note:
    1. Data will be lost if you do not have backup
    2. You must follow step 1 to step 4 VERY CLOSELY.
    3. Repeat the process if necessary.

  • In sm30, when save, how to fill the table with the system data in the table

    Hi all, in SM30, we have user name and time fields, when we create record, after we click the SAVE button, it will automatically fill the fields with the system data, how can we realize it?

    Hi ,
    Do like this.
    1. Go to SE11 -> give your table name and press on change button .
    2.Go to Menu->Utilities-> Table Maintainance Generator , then it will take you another screen .
    3.Assign Functin group if already have one , other wise create function group and assign
    4.Go to Menu->Environment->Modification->Events , again it will take you to another screen
    5.Click on new entries
    ->Now press F4 in T column . Select<u> "01" - Before saving the data in the database</u> and in form routinue give any name of the routinue like "SAVE_USER_DATA" and press enter then u can see an editor symbol in next column click on that it will take you ABAP editor there type this code .
    ztable-username = sy-uname.
    ztable-createtime = sy-uziet.
    check and activate the code .
    With this your job will be complete . Now while entering data just leave columns blank , they will insert corresponding username and time
    Don't forget award points if it helps you.
    Regards,
    Raghav

  • How to retrive the records with in the range

    hi,
    I want to read the records from the table with in the date range.
    for example select empID,empName from EmpTable where stardat to lastdate
    please somebody help me...
    thanx in advance
    cinux

    hi,
    I want to read the records from the table with in the
    date range.
    for example select empID,empName from EmpTable where
    stardat to lastdate
    please somebody help me...
    thanx in advance
    cinuxhi ,
    Oracle provides BETWEEN ..AND.... operator for such requirement.
    SQL> select empno,ename,hiredate from emp where hiredate between to_date('03-dec-80','DD-MON-RR') and sysdate;
    EMPNO ENAME HIREDATE
    7369 SMITH 17-DEC-80
    7499 ALLEN 20-FEB-81
    7521 WARD 22-FEB-81
    7566 JONES 02-APR-81
    7654 MARTIN 28-SEP-81
    7698 BLAKE 01-MAY-81
    7782 CLARK 09-JUN-81
    7788 SCOTT 19-APR-87
    7839 KING 17-NOV-81
    7844 TURNER 08-SEP-81
    7876 ADAMS 23-MAY-87
    EMPNO ENAME HIREDATE
    7900 JAMES 03-DEC-81
    7902 FORD 03-DEC-81
    7934 MILLER 22-FEB-81
    124 johndoe 28-DEC-05
    15 rows selected.
    Hope this may help you.
    Trinath Somanchi,
    Hyderabad.

Maybe you are looking for

  • Why Does My Camera Stop Recording in the Middle of a Video??

    Why does my Rebel T3i stop recording in the middle of me video taping?? Is it because I do not have enough space on my camera for a video? It lets me take plenty more pictures. And when it stops in the middle of a recording, I start recording again,

  • Cell service working on open box Verizon iPad without setting it up

    Bought an open box Verizon iPad 3.  Went through set up but not enter an Id or credit card for cell service.  But cell service is working, and I can't log into the account with any of my ids.  Any thoughts on what's going on?

  • MDX -Children count function Not working in Case statement

    Hi, I am trying to create set when you slice with the Hierarchy member is leaf level , I want a output only that Leaf level . and When I slice with  the parent level , it has to give all the members below that parent level. But the problem here is wh

  • Why are the albums in iphoto 09 not showing up when I upload them to facebook

    I have downloaded photos off my camera and have put them into a new album. When I click on that album they are there. When I go to upload them onto my facebook and other sites the albums are not there. I can not figure out where they are. They are li

  • Update wierdness...

    When trying to update CC applications my updates fail and I get the following error codes (any thoughts will be appreciated): Adobe Premiere Pro CC 2014.0.1 Update There was an error installing this update. Please quit and try again later. Error Code