Multi-Header, Detail and Trailer Record

Hi,
I have to plants, For Each plant i need Header, no.of Details and one trailer Record i need.
Like
Plant 1000
Header1
Deatail Record1
Deatail Record2
Deatail Record3
Trailer1
for plant 2000
Header2
Deatail Record1
Deatail Record2
Deatail Record3
Trailer2
Thanks!
Regards,
SReddy

Hi,
If you could give us the source and target structure with occurrences, it could be easy for us to suggest.
Best regards,
raj.

Similar Messages

  • How to Seggragate Header, Footer and Trailer Records in Informatica Cloud ?

    Hi All, This is my source file Structure which is Flat File. Source File : "RecordType","Creation Date","Interface Name""H","06-08-2015","SFC02""RecordType","Account Number","Payoff Amount","Good Through Date""D","123456787","2356.14","06-08-2015""D","12347","2356.14","06-08-2015""D","123487","235.14","06-08-2015""RecordType","Creation Date","TotalRecordCount""T","06-08-2015","5" The Source File has to be loaded into three targets for Header , Detail and Trailer records separately. Target Files: File 1 : Header.txt
    "RecordType","Creation Date","Interface Name""H","06-08-2015","SFC02" File 2 : Detail.txt "RecordType","Account Number","Payoff Amount","Good Through Date""D","123456787","2356.14","06-08-2015""D","12347","2356.14","06-08-2015""D","123487","235.14","06-08-2015" File 3 : Trailer.txt "RecordType","Creation Date","TotalRecordCount""T","06-08-2015","5"  I tired this solution below :  1.  Source ---> Expression ]-----filter 1---Detail.txt                                        -----filter 2---Trailer.txt  In source , I will read the records starting from 3rd row. This is because, if i read from first row, the detail part contains more fields when compared to Header part. Header Part contains only three fields. So it is taking only first three fields records in Detail Section as well.That's why I am skipping the first two records(Header fields and header record) .. refer the example.. In filter 1, condition is Record_Type = 'D'. In Filter 2 , condition is Record_Type = 'T'.So, the filter 1 will load to Detail.txt and Filter 2 Will load to Trailer.txt In task , pre session command, Calling the windows .bat script to fetch the first two lines and load into Header.txt  This solution is working fine..  My query is can we use two pipeline flow in a same mapping in Informatica cloud.?  Pipeline Flow 1  Source ---> Expression ]-----filter 1---Detail.txt                                    -----filter 2---Trailer.txtPipeline Flow 2  Source ---> Expression ]-----filter 3(Record_Type='H')---Header.txt Source file is same in two flows. In first flow, I will read from the third row, skipping the header section as I mentioned earlier.In second flow , I ll read the entire content and take only header record and load into target.  If I add the flow 2 to existing flow 1.  I am getting the below error..TE_7020 Internal error. The Source Qualifier [Header_Source] contains an unbound field [Creation_Date]. Contact Informatica Global Customer Support.  1. Do informatica Cloud supports, two parallel flow in a same mapping ?2. Any other best solution for my requirement?   Since I am new to Informatica Cloud, Can anyone suggest any other solution if you have?? It will be more helpful if you guys suggest a good solution ..  ThanksSindhu Ravindran

    We are using a Webservices Consumer Tranformation in our mapping to connect to RightFax Server using a WSDL url via Business Service in a mapping.Here in the mapping, where we are sending the input parameters through a flat file and then connecting to Rightfax Server via WS Consumer transformation and then fetching the data and writing to a Flat File.  07/28/2015 10:10:49 **** Importing Connection: Conn_000A7T0B0000000000EM ...07/28/2015 10:10:49 **** Importing Source Definition: SRC_RightFax_txt ...07/28/2015 10:10:49 **** Importing Target Definition: GetFax_txt ...07/28/2015 10:10:49 **** Importing Target Definition: FaultGroup_txt ...07/28/2015 10:10:49 **** Importing SessionConfig: default_session_config ...    <Warning> :  The Error Log DB Connection value should have Relational: as the prefix.    <Warning> :  Invalid value  for attribute Error Log DB Connection. Will use the default value     Validating Source Definition  SRC_RightFax_txt...    Validating Target Definition  FaultGroup_txt...    Validating Target Definition  GetFax_txt...07/28/2015 10:10:49 **** Importing Mapping: Mapping0 ...    <Warning> :  transformation: RIghtfax - Invalid value  for attribute Output is repeatable. Will use the default value Never    [transformation< RIghtfax > ] Validating transformations of mapping Mapping0...Validating mapping variable(s).07/28/2015 10:10:50 **** Importing Workflow: wf_mtt_000A7T0Z00000000001M ...    <Warning> :  Invalid value Optimize throughout for attribute Concurrent read partitioning. Will use the default value Optimize throughput   [Session< s_mtt_000A7T0Z00000000001M >  --> File Reader< File Reader > ]     <Warning> :  The value entered is not a valid integer.    <Warning> :  Invalid value NO for attribute Fail task after wait time. Will use the default value Successfully extracted session instance [s_mtt_000A7T0Z00000000001M].  Starting repository sequence id is [1048287470] Kindly provide us a solution. Attached are logs for reference.

  • Fixed width file outgoing with header,detail and trailer in one map

    we have a scenario where in we have to create an output file. the file is fixed width with three different kinds of records namely header, details and trailer. if the specification of the metadata of the file is specified how will we incorporate the creation of the file in one map since the map produces only one call to UTL_FILE and writes it.
    there is a work around in a crude way where in we can define the file as a CSV and pass everything in an appanded string in one column. but i wanted to know how we can get the three logical records in one map..
    thanks
    kamal

    The problem is not with generating the records.. i mean header , detail and trailer.
    it is how we put them into a fixed width sampled file. if we look at the code generated we will see only one call being made to generate the file.

  • SFTP Sender - content conversion to read Header & detail and control record

    Hi Experts,
        I'm just beginner in SAP PI. Just would like to know possibility for below requirement for my further reseach
    Soure file format fix length  with 3 different format(H,D,T)
    As example
    H001TestHeader1
    D001001TestDetail11
    D001002TestDetail21
    T2
    H002TestHeader2
    D002001TestDetail21
    T2
    The requirement detail :
    1. There are multiple set of H,D,T per files
    2.  T line is the record count of total D line per each H set.  for example as green font we have 2 D line, So T show T2
    3.  If number in T line is not match with actual D line, then only these set should be filter out and write to error file and send email alert. for example as red font we have 1 D line, but T show T2
    4.  If number in T line is match with actual D line, then these set will be write to result file.
    As I know that content  conversion may be support this requirement for separate records set for H/D/T but I don't have idea how to handle requirement no.3 and 4, Is it need ccBPM? Please help to provide some overview guideline for my further research. Thank you.

    As I know that content  conversion may be support this requirement for separate records set for H/D/T
    >>> For the  FCC, you might be using Message Transformation bean and unfortunately MTB doesn't provide the option to split the message i.e. " Records set per message".
    If you go for two step approach i.e.  then you can use this option to split the records.
    1) pass through interface -> SFTP -> NFS
    2) NFS(Standard FCC) -> Your target system
    I don't have idea how to handle requirement no.3 and 4, Is it need ccBPM? Please help to provide some overview guideline for my further research. Thank you.
    >>> If you can split the file for each recordset i.e.  one set of H/D/T then you can handle the validation part in message mapping.
    If T has valid number of records then execute the mapping set the filename  and directory to "result file" locations using dynamic configuration.
    If T has invalid number of records then execute the mapping set the filename  and directory to "error file" location.
    To trigger the email from the message mapping refer to the below blogs -
    Mail without email adapter? Part - I - Process Integration - SCN Wiki
    How to send Alerts using UDF or JAVA Mapping

  • Open Hub Header and Trailer Record.

    Hi,
    For the Open Hub Destination
    Destination Type is File,
    Application server.
    Type of File Name: 2 Logical File name.
    How to to the get the Header and Trailer Record which will contain Creation Date, Creation Time and Total number of records.
    Header record Layput :
    Creation Date (YYYYMMDD)
    Creation Time (HHMMSS)
    Total record Count
    Trailer record Layout:
    Total number of Records in the file and
    XXXX( Key Figure ) Total.
    Thanks in advance.
    Regards,
    Adhvi.

    Hi Venkat,
    write a UDF in following way...
    pass the first parameter as the detail node (cache the whole queue) to the UDF pass the second parameter as the trailer countto the UDF
    now loop through the detail records get the count with a counter variable
    check the counter against the trailer count outside the loop
    if it doesnot match trigger the alert from the UDF itself
    Check the below link for triggering alert from an UDF
    /people/bhavesh.kantilal/blog/2006/07/25/triggering-xi-alerts-from-a-user-defined-function

  • Header and Trailer record validations

    Hi
    I have scenario file-xi-proxy. File contains the header record, detail record and trailer record
    In header record i am getting date field, in header i have to do validation like header record exits and it should be in date format.
    In Trailer record i have to do Total records and Total amount  equals to Total records processed and total amount.
    And amount value should be greater than zero..
    My source structure
    DT_ ACEAwardInformation
    <b>Header</b>
    BeginDate
    EndDate
    <b>DetailRecord</b>
    Field1
    Field2
    Field3
    Field4
    <b>TrailerRecord</b>
    TotalRecord
    TotalAmount
    Even content conversion parameters.
    venkat

    Hi Venkat,
    write a UDF in following way...
    pass the first parameter as the detail node (cache the whole queue) to the UDF pass the second parameter as the trailer countto the UDF
    now loop through the detail records get the count with a counter variable
    check the counter against the trailer count outside the loop
    if it doesnot match trigger the alert from the UDF itself
    Check the below link for triggering alert from an UDF
    /people/bhavesh.kantilal/blog/2006/07/25/triggering-xi-alerts-from-a-user-defined-function

  • Header and Trailer record handling in ODI

    Hi All,
    I have a delimited file which I am trying to load into Oracle DB using Oracle Data Integrator. We are using SQLLDR LKM for this. Since we have header, trailer have different format than the data records, so we have another interface which loads data. Since we have the requirement to load the data into another control table for audit purpose we have another interface.
    What happens when we load the file, because of two interfaces the file scan happens 2 twice. This increases the overall turnaround time. Please note these interfaces work on filter basis for a field called record_type. 0,1,2 are the values for this field which mean header, data and trailer respectively. What we are looking for is how to restrict this double scanning of file for header-trailer and data records. What are the options which can be used here.
    Regards,
    Prashant

    Hi Prashant
    You can use a combination of external table and multi table insert. In ODI 11g there is a Oracle Multi Table Insert IKM. You will have to have one interface which has the source as the file and the target is a temporary target. The LKM will use LKM File to External Table, the IKM will use Oracle Multi Table IKM, and it will set DEFINE_QUERY. Also the LKM should set DELETE_TEMPORARY_OBJECTS to false. Then define 2 interfaces to load your 2 targets for example. The first target loading interface will use the 1st interface as a source and the MTI IKM again will use the MTI IKM and set IS_TARGET_TABLE to true and execute to false. Then the second target loading interface will use the 1st interface as a source and the MTI IKM again wil set IS_TARGET_TABLE to true and execute to true.
    Cheers
    David

  • How to define a schema for fixed length consisting of multiple records of different types having header detail and footer at positions 22 and 23 namely having 10 for header 20 for detail and 30 for footer

    how to define a schema for fixed length consisting of multiple records of different types having header detail and footer at positions 22 and 23 namely having 10 for header ,20 for detail and 30 for footer.
    here is the sample file
    DDWTYBILL13092502572 1000000112\\NHLSFS\ORACLE\ORACLE\BILLING\VERIFY\            2013/09/25 01:19:08 BILLEND     
    DDWTYBILL13092502572 2000000212PRIVATE             PRIVATE             STY5572142     010700         INV                              12STYD000789                  02-AUG-20130000010472      STY       STY       1100611006MGIT                91  NNM    DIMPHANA,CN,CELESTE NERINA                                                                                                        09-JAN-1964                              TYG       A2        DR DU PREEZ                                                                                         5602181             63301345       TB731543                                           Suspect TB ?CA, (L) EMPIEEM                                           Ward A2                                           TYGERBERG HOSPITAL                                                                                  Z03.0                                                       Z03.9                                                          05-AUG-201305-AUG-201324-SEP-2013FEND    113.60
    DDWTYBILL13092502572 2000000312PRIVATE             PRIVATE             STY5572149     010700         INV                              12STYD000789                  02-AUG-20130000010472      STY       STY       1100611006MGIT                91  NNM    DIMPHANA,CN,CELESTE NERINA                                                                                                        09-JAN-1964                              TYG       A2        DR DU PREEZ                                                                                         5602181             63301345       TB731548                                           Empyema BPF                                                           Ward A2                                           TYGERBERG HOSPITAL                                                                                  Z03.9                                                       Z03.9                                                          05-AUG-201305-AUG-201324-SEP-2013FEND    113.60
    DDWTYBILL13092502572 2000000412PRIVATE             PRIVATE             STY5572351     010700         INV                              12STYD000789                  02-AUG-20130000010472      STY       STY       1100611006MGIT                91  NNM    DIMPHANA,CN,CELESTE NERINA                                                                                                        09-JAN-1964                              TYG       A2        DR DU PREEZ                                                                                         5602181             63301345       TB731584                                           Suspect TB ?EMPIEEM, ?CA                                              Ward A2                                           TYGERBERG HOSPITAL                                                                                  Z03.0                                                       Z03.9                                                          05-AUG-201305-AUG-201324-SEP-2013FEND    113.60
    DDWTYBILL13092502572 2000000512PRIVATE             PRIVATE             STY5572352     010700         INV                              12STYD000789                  02-AUG-20130000010472      STY       STY       1100611006MGIT                91  NNM    DIMPHANA,CN,CELESTE NERINA                                                                                                        09-JAN-1964                              TYG       A2        DR DU PREEZ                                                                                         5602181             63301345       TB731585                                           Suspect TB ?EMPIEEM, ?CA                                              Ward A2                                           TYGERBERG HOSPITAL                                                                                  Z03.0                                                       Z03.9                                                          05-AUG-201305-AUG-201324-SEP-2013FEND    113.60
    DDWTYBILL13092502572 2000000612MAJOR               MAJOR               STY5577530     010700         INV                              12ZTYG000009                  02-AUG-20130000010472      STY       STY       1100611006MGIT                91  NNS    ZOZI,S,SIYAMBONGA                                                                                                                 27-MAR-2010                              TYG       D3        DR MAHOMED                                                                                          5602181             113799662      TB731663                                           Lymphadenitis                                                         Ward D3                                           TYGERBERG HOSPITAL                                                                                  I88.9                                                       Z03.9                                                       H1 07-AUG-201307-AUG-201324-SEP-2013MEND     58.52
    DDWTYBILL13092502572 2000000712MAJOR               MAJOR               STY5577537     010700         INV                              12ZTYG000009                  06-AUG-20130000010472      STY       STY       1100611006MGIT                91  NNS    DLAMINI,N,NTLANTLA                                                                                                                09-OCT-1976                              TYG       F1        DR WZESE                                                                                            5602181             135524676      TB731666                                           Suspect TB                                                            Ward F1                                           TYGERBERG HOSPITAL                                                                                  Z03.0                                                       Z03.9                                                       H1 07-AUG-201314-AUG-201324-SEP-2013MEND     58.52
    DDWTYBILL13092502572 30003034120000003032END

    You can create an nXSD using Native Format Builder and use lookAhead attribute to read values at specific position and use that as the choice. You can find an example here
    http://docs.oracle.com/cd/E11036_01/integrate.1013/b28994/nfb.htm#BGBBAJFD

  • Re: to create outbound file with multiple header ,detail and trailor in informatica

    Dear Mohan Prakash, Please mentation The Target Which Format you need. Please let me know. Thanks & RegardsKasireddy+966545281845

    hi I want to create a file with muiltple header detail and trailorSample of my file  1F99500094959                    5F99000000999911025F99000000999912025F99000000999913029F99500094959                    1G83341002729803                5G83000000999918020109G83341002729803                1G83910377940                    5G830000009999190201011050003522029G83910377940                    T20110720000000900000006750{  In this file 1 is header and 5 is detail and 9 is trailer and in last T segment is also trailerI have any idea to create seperate 3 files and in lastconcatenate all file with the help of unixbut in my case data is so large (in millions) this approach is not use full please suggest my any ideaHow to create VSAM file

  • Header, main and trailer in alternate pages

    I want to print Header in first page, main in second and trailer in third page. What I'm getting now when I have 3 records is main in first three pages, then main in the next 3 then trailer in last 3 pages. I set all the section to print in first page. Any inputs?

    That thing I can do.
    But, the requirement is like this: The first page is employee header details on portrait, the second page is let's say employee's timesheet that requires landscape layout, then the third page is summary that is also portrait. Now, I need the three sections as header=portrait, main=timesheet and trailer=portrait. So basically, if I have three employees, the second, fifth and 8th page is landscape. But what I'm getting right now is pages 1-3 = header, 4-6 = main and 7-9 is trailer.

  • Checking header and trailer record

    Hi
    Plz can any one guide me writing UDF for checking whether header record exists or not
    i am using node funtion Exits...
    In UDF i wanna check the value of Exists if it is true continue the process if it is false raise an exception
    How we write UDF for that
    venkat

    Hi.....Venkat,
    1.Import ur sender & reciever XSD files.
    2.A user defined function is created before doing the mapping where we write the java code to achieve the looping conditions.
    3.At the time of creating the function we should have two arguments (ex.a & b)(importing parameters) to it.then the java code is written for the conditions.
    4.write ur code.
    5. then using the user defined function further mapping is done.In this way the conditions are achieved during mapping.
    ******************if it helpful points will be rewarded*********************************
    Regards,
    Naresh.K

  • Excluding detail and suppressed records from calculations in crystal xi

    Is there anyway to exclude records from calculations in crystal xi
    The report design contain group level A, group level B and group level C
    In Group level C - the sorting order ensures that the most recent (and in this scenario relevant record) is displayed first - this is the record which is required for the calculations.  This record is then moved into Group Level C header - thus only displaying the first record in this group. However any calculations which take place take into account all the detail records in Group Level C, and as a result the calculations are incorrect.
    Running Totals are not an option as also on this report - in certain scenarios Group Level C may be suppressed altogether - is there anyway to exclude suppressed records from the calculations?
    It is necessary to carry out the following calculations, using only the (unsuppressed) value from Group Header Leve C: Sum, Mean, Median
    Thank you in advance

    if i understand correctly you need to calcuate values based upon your group c
    and the details of the group are in the details section
    the only way you can accuratley calc this is MANUAL RUNNING TOTALS.
    RESET
    The reset formula is placed in a group header report header to reset the summary to zero for each unique record it groups by.
    whileprintingrecords;
    Numbervar  X := 0;
    CALCULATION
    The calculation is placed adjacent to the field or formula that is being calculated.
    (if there are duplicate values; create a group on the field that is being calculated on. If there are not duplicate records, the detail section is used.
    whileprintingrecords;
    Numbervar  X := x + ; ( or formula)
    DISPLAY
    The display is the sum of what is being calculated. This is placed in a group, page or report footer. (generally placed in the group footer of the group header where the reset is placed.)
    whileprintingrecords;
    Numbervar  X;
    X

  • R12 Positive Pay eText Template with File Header/Trailer Records

    A client for whom we are implementing R12.1.3 Positive Pay uses a bank which requires file header and trailer records in the file transmission. We used the standard IBYPOS_PAY_en.rtf template file and modified it to provide the additional, required detail. The new version of the template appears to be working exactly as needed when viewing the output from the template viewer. However, when positive pay is run through eBS there is no text output generated. The XML for report is being generated, as it can be pulled from both the concurrent request log file and the iby_trxn_documents table. Below is a snippet of the pertinent information from the log file:
    After formatting length of the output - baos::0
    After formatting, output is not null
    Output type is eText, so logging the output below:
    ==================================================
    ==================================================
    length of the output - baos::0
    After formatting, periodicSeq is null or empty.nullTimestamp: Fri Jan 11 12:02:53 EST 2013
    length of the output - baos::0
    Exit: iby.scheduler.FDFormatProgramUtils.doExtractAndFormat()Timestamp: Fri Jan 11 12:02:53 EST 2013
    Saving format output to file system.
    Enter: iby.scheduler.FDFormatProgramUtils.saveToFileSystemCustom()
    Concurrent request output file name: /u02/app/appti/TCDXI/inst/apps/TCDXI_ncdx201/logs/appl/conc/out/o455658.out
    File.separator: /
    Custom file name: /u02/app/appti/TCDXI/inst/apps/TCDXI_ncdx201/logs/appl/conc/out/CDX_455658.out
    Opened file in append mode: /u02/app/appti/TCDXI/inst/apps/TCDXI_ncdx201/logs/appl/conc/out/CDX_455658.out
    Updating FND request record for the custom file location/u02/app/appti/TCDXI/inst/apps/TCDXI_ncdx201/logs/appl/conc/out/CDX_455658.out
    request id: 455658
    Exit: iby.scheduler.FDFormatProgramUtils.saveToFileSystemCustom()
    Updating positive pay file created flag on the instruction record.
    Enter: iby.scheduler.FDFormatProgramUtils.postResults()
    Sql to be executed: BEGIN IBY_FD_POST_PICP_PROGS_PVT.set_pos_pay_created(:1); END;
    Payment instruction id: 10001
    Exit: iby.scheduler.FDFormatProgramUtils.postResults()
    Successfully updated instruction record.
    Exit: iby.scheduler.FDPositivePay.runPosPay
    Program completed successfully.
    java.util.logging.ErrorManager: 2:
    oracle.core.ojdl.LoggingException: oracle.core.ojdl.LoggingException: Attempt to flush a closed LogWriter
         at oracle.core.ojdl.ExceptionHandler.onException(Unknown Source)
         at oracle.core.ojdl.BufferedLogWriter.handleException(Unknown Source)
         at oracle.core.ojdl.BufferedLogWriter.flush(Unknown Source)
         at oracle.core.ojdl.logging.ODLHandler.flush(Unknown Source)
         at oracle.core.ojdl.logging.ODLHandler.publish(Unknown Source)
         at java.util.logging.Logger.log(Logger.java:478)
         at java.util.logging.Logger.doLog(Logger.java:500)
         at java.util.logging.Logger.log(Logger.java:523)
         at oracle.ias.cache.CacheInternal.logLifecycleEvent(Unknown Source)
         at oracle.ias.cache.CacheInternal.close(Unknown Source)
         at oracle.ias.cache.Cache.close(Unknown Source)
         at oracle.apps.jtf.cache.IASCacheProvider$CacheStopperThread.run(IASCacheProvider.java:1480)
    Caused by: oracle.core.ojdl.LoggingException: Attempt to flush a closed LogWriter
         ... 10 more
    Start of log messages from FND_FILE
    Enter:TIMESTAMP:: IBY_EXTRACTGEN_PVT.Create_Extract:: 11-JAN-13 12.02.52.760550000 PM -05:00
    Enter:TIMESTAMP::Pos Pay Extract(5) ::11-JAN-13 12.02.52.761687000 PM -05:00
    Enter:TIMESTAMP:: POS Payiby.plsql.IBY_FD_EXTRACT_GEN_PVT.Create_Pos_Pay_Extract_1_0:: 11-JAN-13 12.02.52.761942000 PM -05:00
    Exit:TIMESTAMP:: POS Payiby.plsql.IBY_FD_EXTRACT_GEN_PVT.Create_Pos_Pay_Extract_1_0:: 11-JAN-13 12.02.53.121536000 PM -05:00
    Exit:TIMESTAMP::Pos Pay Extract(5) ::11-JAN-13 12.02.53.121603000 PM -05:00
    End of log messages from FND_FILE
    Executing request completion options...
    Output file size:
    0
    ------------- 1) PRINT   -------------
    Disabling requested Output Post Processing. Nothing to process. The output of the request is zero byte.
    Any help that may be offered is greatly appreciated.
    Regards,
    Paul

    I faced this issue when I ran the program as a stand alone program. Please make sure that the Postive pay file is selected in the Payment Process profile which is used to make the payment. And submit the PPR and then run the positive pay from there(Payments Manager page) by selecting the check box.

  • How to do the header and trailer validation in the input file?

    hi,
    what are the ways we can validate whether header and trailer record exists in the input file?
    how to do that?
    regards
    Ruban

    File to Proxy Validation
    https://www.sdn.sap.com/irj/sdn/go/portal/prtroot/docs/library/uuid/99593f86-0601-0010-059d-d2dd39dceaa0
    /people/swaroopa.vishwanath/blog/2005/06/24/generic-approach-for-validating-incoming-flat-file-in-sap-xi--part-1
    /people/swaroopa.vishwanath/blog/2005/06/29/generic-approach-for-validating-incoming-flat-file-in-sap-xi--part-ii
    /people/prateek.shah/blog/2005/06/14/file-to-r3-via-abap-proxy

  • Refreshing Header & Detail part in Screen

    Dear Friends,
    I've been facing the following two problems in the diaog program that I am developing:
    In my screen I have header & detail portions. The header part contains fields e.g. contract no., transporter code, Date effective from, Date effective upto etc. that I pick up from transparent tables. To display the detail part I am using a table control that gets populated from internal table. Now when I save, the data gets saved successfully (both header & detail) and subsequently I am clearing the screen also but still I am not able to enter a  new entry.There is no alternative other than to come out of the program & executing it once again.
    Further if I change any entry in the header part (containing only fields) and save that updation doesn't get saved  i.e.next time also it shows the previous values while in the detail part (table-control) the updation gets saved.
    Waiting for your invaluable inputs..
    Regards,
    Alok.

    Hi Alok,
    Seems in your program.. After change in details part... while saving the modules written in PBO are retriving old values.
    Please put some flag in PBO which will control it's execution. Ie. at 1st time when u fetch the data. It will set as "X' then for every next time when u change data in details... It will not fetch data. It will accept your own data in header.
    Darshan

Maybe you are looking for