ICR_ FBICR3_ Process 003_ assign vs unassign data records

I'm setting up ICR process 003. I have couple of questions:-
1)     In transaction FBICR03 what is the difference between unassigned data records and assigned data record.
2)     Is it possible to get rid of assigned data records in bottom of the screen so all the records shows up under unassigned data record
Thanks for all the support!

Ralph Stadter can you please advice with above question.
Thanks

Similar Messages

  • SAP MII 14.0 SP5 Patch 11 - Error has occurred while processing data stream Dynamic Query role is not assigned to the Data Server

    Hello All,
    We are using a two tier architecture.
    Our Corp server calls the refinery server.
    Our CORP MII server uses user id abc_user to connect to the refinery data server.
    The user id abc_user has the SAP_xMII_Dynamic_Query role.
    The data server also has the checkbox for allow dynamic query enabled.
    But we are still getting the following error
    Error has occurred while processing data stream
    Dynamic Query role is not assigned to the Data Server; Use query template
    Once we add the SAP_xMII_Dynamic_Query role to the data server everything works fine. Is this feature by design ?
    Thanks,
    Kiran

    Thanks Anushree !!
    I thought that just adding the role to the user and enabling the dynamic query checkbox on the data server should work.
    But we even needed to add the role to the data server.
    Thanks,
    Kiran

  • Assigning today's date to process flow start actvity

    Hi,
    I am trying to assign today's date to my process flow which will be used as a common load date for multiple transformations within the process flow. I'm at a loss on how to do this. Any help would be appreciated.
    Thanks

    You can add variable to the start acitivity or create stand alone variable(global for Process Flow) in the left side panel (where you see selected objects/events) and give the value of the variable as sysdate. Make sure the Literal property is set to false meaning it is not literally SYSDATE but the value of sysdate.
    It worked for me.
    - Manohar

  • Process Instance.Task Details.Update Date

    Hi experts,
    Process Instance.Task Details.Update Date
    Process Instance.Task Information.Update Date
    From these two dates which one captures the approval task assignment date.
    For e.g. If we reassign the approval task after some idle period to another approver which date will give this reassignment date?

    There's also information in the OSH and OSI table.
    Process Instance.Task Information.Assigned Date
    OSI_ASSIGNED_DATE
    Process Instance.Task History.Assign Date
    OSH_ASSIGN_DATE
    You might need to create an approval task and have inserted, and then reassign and do some testing to find the actual value you need to search for.
    Another note, any open task that is Rejected, Waiting, or Pending will be in the OTI table which should contain all these values.
    -Kevin

  • Could not get F4 help assigne to Formula date on web

    Hi All,
    When i run my query in web , i could not get F4 help assigne to Formula date variable which is processing by customer exit and ready for input.
    can u lease help me in this .
    Thanks
    Man

    See these Norton support threads for information about Firefox 8.0 compatibility.
    360 <br />
    http://community.norton.com/t5/Norton-360/Firefox-8-Support-for-Norton-Toolbar/td-p/581640
    NIS <br />
    http://community.norton.com/t5/Norton-Internet-Security-Norton/Firefox-8-Support-for-Norton-Toolbar/td-p/581516

  • ICR Process 2, FBICR2, reconcilation screen not showing data records

    Experts,
    My reconciliation screen is not showing me any records for process 2. After posting 2 documents and running FBICA2 (Assignment Program), I can see the items be assigned. The records are also all available in auxially programs total records and list of documents. When I run FBICR2 (manual reconciliation) I get the following screen http://img20.imageshack.us/img20/8163/20395360.png
    . Has anyone dealt with this issue? Why would the records be available at all steps except the last?

    Ralph,
    With the help of an ICR SME yesterday, I was able to get process two working! It was great, the two notes helped and we reconciled several test scenarios.
    This morning, I was working on process 2 and used the auxilary program to delete the transactional data (not index tables), reactivated the data tables, posted two new documents, and now the reconciliation display is back to u201CNo Data Record Selectedu201D issue we solved this week. No configuration was changed since FBICR2 began working yesterday. I tried recalculating totals in the Aux. programs list but that didn't help either.  FBICA2 is still working perfectly and when I look at the records in auxiliary programs they have been assigned a document group number, so I know they are reconciling. The issue is with the Display in FBICR2. What would trigger FBICR2 to revert back to this error? Are there any known issue with the stability of FBICR2, or known issues with using the delete transactional data program?
    Please Advise.
    Thanks,
    John Scime

  • When I plug in my iPhone to sync, it starts but does not complete the process and several pieces of data are not being transferred to my iPhone from my MacBook Air.

    Problem:
    When I plug in my iPhone to sync, it starts but does not complete the process and several pieces of data are not being transferred to my iPhone from my MacBook Air.
    Any help that you can provide in helping my iPhone accurately sync with iPhoto and iTunes will be greatly appreciated.
    Symptoms:
    1)   Sync:  It’s not completing the sync.  Below, I’ve provided all of my settings from the iTunes Summary tab so that you might tell me if I’ve selected any incorrect options.  I prefer to sync the “old school” way – by connecting to the computer – as opposed to syncing over the cloud. Perhaps this is what’s causing the problem? Here is a list of the steps displayed in the iTunes window as the sync progresses:
    waiting for sync to start (step 1 of 7)
    backing up (step 2 of 7)
    preparing to sync (step 3 of 7)
    determining apps to sync (step 4 of 7)
    preparing apps to sync (step 5 of 7)
    importing photos (step 6 of 7)
    waiting for changes to be applied (step 7 of 7)
    syncing apps / copying 1 of 4 (App name) (step 7 of 7)
    canceling sync
    apple icon
    2)   Photos: I've selected only certain of my iPhoto albums to sync to my iPhone.  All of the albums are correct/complete in iPhoto.  All of the albums are listed on my iPhone, both before and after the sync, but the albums are empty (no actual photos) before and after the sync. Perhaps this is tied to the fact that the sync isn’t completing, but because “importing photos” is one of the steps that the incomplete sync displays, I don’t know.
    3)   Apps: When I launch iTunes and click on the Apps folder under the Library listing, then click on the Updates tab, iTunes searches for any Apps needing to be updated and provides a list.  If I click on Update All, the Apps are successfully updated in iTunes.  But, when I plug in my iPhone so that the updates will transfer to the apps on my iPhone, the updates don’t transfer to the apps on my iPhone and those apps still reflect that they need updating on the iPhone.
    Other Potential Pertinent Info:
    The flash memory hard drive on my MacBook Air recently died (perhaps a month or two ago).  Apple had emailed me about a known issue and literally the next day, my MacBook Air crashed.  I installed a new flash memory drive and re-installed everything from a backup off of an external hard drive.  Everything seems to be working fine; it recreated accurately all of my software and data, including iPhoto and iTunes, the pictures and songs (respectively) for which are stored on that hard drive, as opposed to being on the flash memory in the MacBook Air itself.  However, I don’t recall if the start of the sync problem described herein started happening at the same time that I replaced the flash memory drive.  All I know is that the computer is working perfectly in all respects and that even as the sync is failing, it at least says that it’s doing the right things and looking in the right places (e.g., the list of albums on my iPhone matches the list of albums in iTunes, etc.).
    Settings/Status:
    MacBook Air
    OSX v. 10.9
    iPhoto ’11 v. 9.5 (902.7)
    iPhone iOS 7.0.4
    iTunes v. 11.1.3 (8)
    Summary Tab
    Backups (This Computer)
    Options
    Automatically sync when this iPhone is connected
    Sync only checked songs and videos
    Photos Tab
    Sync Photos from iPhoto (429 Photos)
    Selected albums, Events, and Faces, and automatically include (no Events)
    Albums – 9 are selected

    You need to download iTunes on your computer. iOS 6 requires the latest version of iTunes which is 10.7.

  • How does the GR processing time affect the scheduling of the process order & the latest start date in the operation.

    Hi
    Can anyone explain  how does the GR processing time affect the scheduling of the process order & the latest start date in the operation overview.

    Hi
    GR processing time means number of workdays required after receiving the material in storage.
    Check this link:GR Processing time
    Regards,
    Anupam Sharma

  • GL assignment for future date?

    Hi everybody,
    We have an issue is as follows:
    - We have created a sales order for 08-09-2008
    - We have created PO for 08-09-2008, here system will able to post everything fine.
    - Again we are creating one more PO for 08-09-2009, here system threws error message like GL a/c needs cost assignment.
    My question is why system is asking assignment for future date only and why system is not asking assignment for current date?
    Looking forward your best thoughts
    regards
    vk

    Hi Raj,
    Thanks for your reply
    Yes, we have assigned only cost element in TC: OKB9. But, for that cost element we have created 2 analysis periods.
    - GL 100115            01.01.2008 to 31.12.2008
    - GL 100115            01.01.2009 to 31.12.9999
    I think validity period is perfect still system threws error message like gl acc needs cost object assign.
    regards
    vk

  • Error processing request  ORACLE-01403:no data found

    We surmounted the blank page issue by applying the patch to APEX 4.2.2.
    Now we patched/upgraded to APEX 4.2.3, and no longer have the blank page issue, but have this one that seems similar:
    I am trying here, since I am not getting any debug messages - the POST goes in and immediately comes back with the no data found error.
    APEX 4.2.3
    APEX Listener 2.0.2
    Glassfish 4.0
    Apps install OK, and can display the Login page (Show), but when trying to Login (Accept),
    immediately get this error:
    Error Error processing request
              ORA-010403: no data found
    Debug messages only shows Show entries - nothing , not even the Accept first line.
    Firebug shows similar - upon Post (which looks OK) the immediate response is the error page and message.
    I do not have access to the APEX Listener logs tonight,  but expect to get to them, or have someone check them
    out, tomorrow AM.
    Any thoughts or suggestions?
    This happens across all apps (15) except for one.  Am still trying to discern the different in that app vs the others.
    Thank you -

    An update:
    I have verified that some apps work, and some apps show the
    Error Error processing request
            ORA-01403: no data found
    error.
    Happens across different workspaces.
    Have confirmed that for an app that has the error,
       I get the error using the APEX Listener deployment. 
       I do NOT get the error using the HTTP server deployment.
    This is the same in a Linux environment and in a Windows environment.
    For one app, I have an earlier version that gives the error, a later version (new app #) that does not give the error.
    APEX 4.2.3
    APEX Listener 2.0.3.221.10/3
    Glassfish 4.0
    Again even at  Debug Level 9, I get no debug entries for the POST
    For an app that gives the error, the APEX Listener log entries are (from the POST):
    ==== Processing Request: ====
    Attempting to process with PL/SQL Gateway
    user-agent: Mozilla/5.0 (Windows NT 5.1; rv:23.0) Gecko/20100101 Firefox/23.0
    host: qdcls1534:8082
    Applied database connection info
    POST /apex/apexd2/wwv_flow.accept
    ==== Headers in Request ====
    ==== Cookies in Request ====
    content-length: 412
    content-type: application/x-www-form-urlencoded;charset=UTF-8
    request parameter: p_flow_id=440
    request parameter: p_flow_id=440
    request parameter: p_flow_step_id=101
    request parameter: p_instance=1712832201985
    request parameter: p_instance=1712832201985
    request parameter: p_page_submission_id=8493461183923
    request parameter: p_page_submission_id=35145982114211
    request parameter: p_request=LOGIN
    request parameter: p_request=
    request parameter: p_debug=LEVEL9
    request parameter: p_debug=LEVEL9
    request parameter: p_arg_names=13616030313099104709
    request parameter: p_arg_names=13616030522289104711
    request parameter: p_t01=karen.x.cannell
    request parameter: p_t02=sssssssss
    request parameter: p_md5_checksum=
    request parameter: p_page_checksum=C2302553B43577579FF77585749CA016
    Using Procedure:wwv_flow.accept
    request parameter: p_flow_step_id=101
    Requesting Pool:apexd2
    pool exists: apexd2
    isValidRequest(), procedure name: <wwv_flow.accept>
    Validating: wwv_flow.accept
    *** Total number of arguments: 510
    SID: 214
    *** Total number of arguments: 510
    begin
    wwv_flow.accept(p_flow_step_id=>?,
    p_md5_checksum=>?,
    p_arg_names=>?,
    p_page_checksum=>?,
    p_t02=>?,
    p_t01=>?,
    p_debug=>?,
    p_request=>?,
    p_page_submission_id=>?,
    p_flow_id=>?,
    p_instance=>?);
    commit;
      end;"
    Parse: 0 ms
    p_flow_step_id= null
    p_flow_step_id=[101, 101]
    {p_md5_checksum=, p_page_submission_id=[8493461183923, 35145982114211], p_debug=[LEVEL9, LEVEL9], p_page_checksum=C2302553B43577579FF77585749CA016, p_request=[LOGIN, ], p_t02=sssssssss, p_t01=karen.x.cannell, p_flow_id=[440, 440], p_instance=[1712832201985, 1712832201985], p_flow_step_id=[101, 101], p_arg_names=[13616030313099104709, 13616030522289104711]}
    p_request=[LOGIN, ]
    p_md5_checksum=
    {p_md5_checksum=, p_page_submission_id=[8493461183923, 35145982114211], p_debug=[LEVEL9, LEVEL9], p_page_checksum=C2302553B43577579FF77585749CA016, p_request=[LOGIN, ], p_t02=sssssssss, p_t01=karen.x.cannell, p_flow_id=[440, 440], p_instance=[1712832201985, 1712832201985], p_flow_step_id=[101, 101], p_arg_names=[13616030313099104709, 13616030522289104711]}
    p_arg_names=[13616030313099104709, 13616030522289104711]
    p_arg_names: {13616030313099104709, 13616030522289104711}
    {p_md5_checksum=, p_page_submission_id=[8493461183923, 35145982114211], p_debug=[LEVEL9, LEVEL9], p_page_checksum=C2302553B43577579FF77585749CA016, p_request=[LOGIN, ], p_t02=sssssssss, p_t01=karen.x.cannell, p_flow_id=[440, 440], p_instance=[1712832201985, 1712832201985], p_flow_step_id=[101, 101], p_arg_names=[13616030313099104709, 13616030522289104711]}
    {p_md5_checksum=, p_page_submission_id=[8493461183923, 35145982114211], p_debug=[LEVEL9, LEVEL9], p_page_checksum=C2302553B43577579FF77585749CA016, p_request=[LOGIN, ], p_t02=sssssssss, p_t01=karen.x.cannell, p_flow_id=[440, 440], p_instance=[1712832201985, 1712832201985], p_flow_step_id=[101, 101], p_arg_names=[13616030313099104709, 13616030522289104711]}
    p_instance= null
    p_instance=[1712832201985, 1712832201985]
    {p_md5_checksum=, p_page_submission_id=[8493461183923, 35145982114211], p_debug=[LEVEL9, LEVEL9], p_page_checksum=C2302553B43577579FF77585749CA016, p_request=[LOGIN, ], p_t02=sssssssss, p_t01=karen.x.cannell, p_flow_id=[440, 440], p_instance=[1712832201985, 1712832201985], p_flow_step_id=[101, 101], p_arg_names=[13616030313099104709, 13616030522289104711]}
    p_flow_id= null
    p_flow_id=[440, 440]
    {p_md5_checksum=, p_page_submission_id=[8493461183923, 35145982114211], p_debug=[LEVEL9, LEVEL9], p_page_checksum=C2302553B43577579FF77585749CA016, p_request=[LOGIN, ], p_t02=sssssssss, p_t01=karen.x.cannell, p_flow_id=[440, 440], p_instance=[1712832201985, 1712832201985], p_flow_step_id=[101, 101], p_arg_names=[13616030313099104709, 13616030522289104711]}
    p_page_submission_id= null
    p_page_submission_id=[8493461183923, 35145982114211]
    {p_md5_checksum=, p_page_submission_id=[8493461183923, 35145982114211], p_debug=[LEVEL9, LEVEL9], p_page_checksum=C2302553B43577579FF77585749CA016, p_request=[LOGIN, ], p_t02=sssssssss, p_t01=karen.x.cannell, p_flow_id=[440, 440], p_instance=[1712832201985, 1712832201985], p_flow_step_id=[101, 101], p_arg_names=[13616030313099104709, 13616030522289104711]}
    p_request= null
    p_md5_checksum=
    {p_md5_checksum=, p_page_submission_id=[8493461183923, 35145982114211], p_debug=[LEVEL9, LEVEL9], p_page_checksum=C2302553B43577579FF77585749CA016, p_request=[LOGIN, ], p_t02=sssssssss, p_t01=karen.x.cannell, p_flow_id=[440, 440], p_instance=[1712832201985, 1712832201985], p_flow_step_id=[101, 101], p_arg_names=[13616030313099104709, 13616030522289104711]}
    p_debug= null
    p_debug=[LEVEL9, LEVEL9]
    {p_md5_checksum=, p_page_submission_id=[8493461183923, 35145982114211], p_debug=[LEVEL9, LEVEL9], p_page_checksum=C2302553B43577579FF77585749CA016, p_request=[LOGIN, ], p_t02=sssssssss, p_t01=karen.x.cannell, p_flow_id=[440, 440], p_instance=[1712832201985, 1712832201985], p_flow_step_id=[101, 101], p_arg_names=[13616030313099104709, 13616030522289104711]}
    p_t01= karen.x.cannell
    p_t01=karen.x.cannell
    {p_md5_checksum=, p_page_submission_id=[8493461183923, 35145982114211], p_debug=[LEVEL9, LEVEL9], p_page_checksum=C2302553B43577579FF77585749CA016, p_request=[LOGIN, ], p_t02=sssssssss, p_t01=karen.x.cannell, p_flow_id=[440, 440], p_instance=[1712832201985, 1712832201985], p_flow_step_id=[101, 101], p_arg_names=[13616030313099104709, 13616030522289104711]}
    p_t02= sssssssss
    p_t02=sssssssss
    {p_md5_checksum=, p_page_submission_id=[8493461183923, 35145982114211], p_debug=[LEVEL9, LEVEL9], p_page_checksum=C2302553B43577579FF77585749CA016, p_request=[LOGIN, ], p_t02=sssssssss, p_t01=karen.x.cannell, p_flow_id=[440, 440], p_instance=[1712832201985, 1712832201985], p_flow_step_id=[101, 101], p_arg_names=[13616030313099104709, 13616030522289104711]}
    p_page_checksum= C2302553B43577579FF77585749CA016
    p_page_checksum=C2302553B43577579FF77585749CA016
    Exec: 37 ms
    ==== Headers from Results ====
    Setting Content-Type (Content-type): text/html; charset=UTF-8
    Got results length: 307
    Processed PL/SQL Gateway request
    ==== Request Processed ====
    It just stops ...
    Any ideas?
    I have tried all of the Compatibility settings, no difference.
    If it is something I can change in the app, that would be great.
    The app that does not work was imported from an APEX 4.1 instance.
    The app that does work was imported from an APEX 4..3 instance (Windows, APEX 4.2.3, APEX Listener 2.0.3, recently upgraded from APEX 4.2.2 and APEX Listener 2.0.2).
    Any help or suggestions will be greatly appreciated,
    Karen

  • Error processing request  ORACLE-01403:no data found (pt.2)

    Hi, all.
    We've experienced the problem that seems to be resolved in topic
    Error processing request  ORACLE-01403:no data found
    I followed the lead but didn't found problem they were talking about - no two #FORM_OPEN# directive in page template.
    We have Apex 4.2.2.00.11 patchset installed
    I cannot tell Apex Listner version for now because our sysadmin who installed it is absent
    And problem appears in different versions of Internet Expolrer, but application is working fine in Firefox

    Hi, all.
    We've experienced the problem that seems to be resolved in topic
    Error processing request  ORACLE-01403:no data found
    I followed the lead but didn't found problem they were talking about - no two #FORM_OPEN# directive in page template.
    We have Apex 4.2.2.00.11 patchset installed
    I cannot tell Apex Listner version for now because our sysadmin who installed it is absent
    And problem appears in different versions of Internet Expolrer, but application is working fine in Firefox

  • Error in process chain when running package (data manager)

    Hello experts,
    I try to run the following how to : "move date between applications using Data Manager ..."
    Yon can find the pdf here : http://www.sdn.sap.com/irj/scn/index?rid=/library/uuid/b0480970-894f-2d10-f9a5-d4b1160be203
    I'm looking after someone who used it successfully, on my side after implementing all steps and running the package I get the message "IMMEDIATE RUN: Request to run the package on the server was successful The package is now running"
    When I look the log for the process chain, overall status of the process chain is yellow, and steps "modify dynamically" and "clear BPC Tables" are in error.
    I can't get any usefull error messages.
    If someone has an idea to have details about those errors, it will be appreciated.
    Points will be awarded, thanks in advance.
    Guillaume P.

    Hello Experts,
    I am also facing the same issue with PC failing at Modify_Dynamically step....
    This process type is working fine when used  in standard delivered process chain to load transaction data.
    I tried using this variant in my cutsom chain, but it fails.
    Please suggest.....
    Thx....D

  • SMS_DISCOVERY_DATA_MANAGER Message ID 2636 and 620. Discovery Data Manager failed to process the discovery data record (DDR)

    Hi
    I'm seeing this critical error on my primary.
    SMS_DISCOVERY_DATA_MANAGER Message ID 2636 and 620. 
    Discovery data manager failed to process the discovery data record (DDR)"D:\Prog.....\inboxes\auth\ddm.box\userddrsonly\adu650dh.DDR", because it cannot update the data source.
    Where these ddr's are actually under ddm.box\userddrsonly\BAD_DDRS folder
    I see a ton of DDR files in that folder. Not sure if I can delete them, so I moved them to a temp folder. AD User discovery keeps generating them.
    Any help ?
    Thanks
    UK
    

    Check the ddm.log file for more information.
    My Blog: http://www.petervanderwoude.nl/
    Follow me on twitter: pvanderwoude

  • Using FIFO as sql code for assigning indicator to data items

    Hi All,
    We are looking for a solution based on FIFO algorithm. We have a table having following data:
    We need to perform FIFO on this table, and assign "object" as data items to other rows based on following conditions:
    1. first we have to group the rows based on "object" column
    2. then we have to traverse each group from the row having  minimum start time of the object group e.g: row id 1 for object group for "19O"
    2.1 Assign a "EqpREf" as "object" + <an integer value> , where integer value changes when the start and end chain finishes. Start -end chain is explained in step 2.2
    2.2 then we have to pick the "nextstarttime" of the picked row and compare it against the closest "starttime" of the rows having "start" as same as "end" of the same row we are
    picking for "nextstarttime" e.g: row id 2 of object 19O is having "nextstarttime" 0310 closest to "starttime" 0355 of row id 2 of object 19O , and rowid 2 is having "start" AAL which is similar to "end" of
    rowid 1 
    2.3 We have to perform this chain till we find end of same and allocate each row in a chain same "Eqpref"
    hence the output we need to generate will come as:
    Kindly help on same.
    Thanks in advance
    -Regards
    Kumud

    Hi,
    Please find the following code block for what is input data and what should be output.
    --The input data
    create table temp_table
    row_id int,
    engine_no varchar(20),
    schedule_no varchar(20),
    start_station varchar(20),
    end_station varchar(20),
    startdate datetime,
    enddate datetime,
    starttime datetime,
    endtime datetime,
    record_id int,
    engine_id int,
    Mgt int,
    nextstarttime datetime,
    Schedule_ref varchar(20),
    Engine_Ref varchar(20)
    GO
    insert into temp_table values(1,'19O','101','SGC','IXP','2015/01/01','2015/01/01','00:00:00','01:00:00',1,10,60,'02:00:00',null,null)
    insert into temp_table values(2,'19O','102','SGC','IXP','2015/01/01','2015/01/01','00:30:00','01:30:00',2,10,60,'02:30:00',null,null)
    insert into temp_table values(3,'19O','103','SGC','IXP','2015/01/01','2015/01/01','02:30:00','03:30:00',3,10,60,'04:30:00',null,null)
    insert into temp_table values(4,'19O','104','IXP','DFW','2015/01/01','2015/01/01','03:30:00','04:00:00',4,10,60,'05:00:00',null,null)
    insert into temp_table values(5,'19O','105','IXP','DFW','2015/01/01','2015/01/01','04:30:00','05:30:00',5,10,60,'06:30:00',null,null)
    insert into temp_table values(6,'19O','106','DFW','ABC','2015/01/01','2015/01/01','05:00:00','06:00:00',6,10,60,'07:00:00',null,null)
    insert into temp_table values(7,'19O','107','DFW','ABC','2015/01/01','2015/01/01','06:00:00','07:00:00',7,10,60,'08:00:00',null,null)
    insert into temp_table values(8,'19O','108','DFW','ABC','2015/01/01','2015/01/01','07:00:00','08:00:00',8,10,60,'09:00:00',null,null)
    insert into temp_table values(9,'19O','109','ABC','DEF','2015/01/01','2015/01/01','10:00:00','11:30:00',9,10,60,'12:30:00',null,null)
    insert into temp_table values(10,'19O','110','XYZ','BDW','2015/01/01','2015/01/01','13:00:00','15:00:00',10,10,60,'16:00:00',null,null)
    insert into temp_table values(1,'319','111','PQR','STU','2015/01/01','2015/01/01','00:00:00','01:00:00',11,11,60,'02:00:00',null,null)
    insert into temp_table values(2,'319','211','PQR','STU','2015/01/01','2015/01/01','04:30:00','15:30:00',12,11,60,'16:30:00',null,null)
    insert into temp_table values(3,'319','112','STU','ABC','2015/01/01','2015/01/01','05:00:00','06:00:00',13,11,60,'07:00:00',null,null)
    insert into temp_table values(4,'319','212','STU','DEF','2015/01/01','2015/01/01','06:00:00','07:00:00',14,11,60,'08:00:00',null,null)
    insert into temp_table values(5,'319','213','STU','PQR','2015/01/01','2015/01/01','07:00:00','08:00:00',15,11,60,'09:00:00',null,null)
    insert into temp_table values(6,'319','118','STU','XYZ','2015/01/01','2015/01/01','10:00:00','11:30:00',16,11,60,'12:30:00',null,null)
    insert into temp_table values(7,'319','119','DEF','JKL','2015/01/01','2015/01/01','13:00:00','15:00:00',17,11,60,'16:00:00',null,null)
    insert into temp_table values(8,'319','215','ABC','MNO','2015/01/01','2015/01/01','17:00:00','20:00:00',18,11,60,'21:00:00',null,null)
    insert into temp_table values(1,'19O','101','SGC','IXP','2015/01/01','2015/01/01','00:00:00','01:00:00',1,10,60,'02:00:00',null,null)
    insert into temp_table values(2,'19O','102','SGC','IXP','2015/01/01','2015/01/01','00:30:00','01:30:00',2,10,60,'02:30:00',null,null)
    insert into temp_table values(3,'19O','103','SGC','IXP','2015/01/01','2015/01/01','02:30:00','03:30:00',3,10,60,'04:30:00',null,null)
    insert into temp_table values(4,'19O','104','IXP','DFW','2015/01/01','2015/01/01','03:30:00','04:00:00',4,10,60,'05:00:00',null,null)
    insert into temp_table values(5,'19O','105','IXP','DFW','2015/01/01','2015/01/01','04:30:00','05:30:00',5,10,60,'06:30:00',null,null)
    insert into temp_table values(6,'19O','106','DFW','ABC','2015/01/01','2015/01/01','05:00:00','06:00:00',6,10,60,'07:00:00',null,null)
    insert into temp_table values(8,'19O','108','DFW','ABC','2015/01/01','2015/01/01','07:00:00','08:00:00',8,10,60,'09:00:00',null,null)
    insert into temp_table values(9,'19O','109','ABC','DEF','2015/01/01','2015/01/01','10:00:00','11:30:00',9,10,60,'12:30:00',null,null)
    insert into temp_table values(1,'319','111','PQR','STU','2015/01/01','2015/01/01','00:00:00','01:00:00',11,11,60,'02:00:00',null,null)
    insert into temp_table values(2,'319','211','PQR','STU','2015/01/01','2015/01/01','04:30:00','15:30:00',12,11,60,'16:30:00',null,null)
    insert into temp_table values(3,'319','112','STU','ABC','2015/01/01','2015/01/01','05:00:00','06:00:00',13,11,60,'07:00:00',null,null)
    insert into temp_table values(4,'319','212','STU','DEF','2015/01/01','2015/01/01','06:00:00','07:00:00',14,11,60,'08:00:00',null,null)
    insert into temp_table values(5,'319','213','STU','PQR','2015/01/01','2015/01/01','07:00:00','08:00:00',15,11,60,'09:00:00',null,null)
    insert into temp_table values(6,'319','118','STU','XYZ','2015/01/01','2015/01/01','10:00:00','11:30:00',16,11,60,'12:30:00',null,null)
    insert into temp_table values(7,'319','119','DEF','JKL','2015/01/01','2015/01/01','13:00:00','15:00:00',17,11,60,'16:00:00',null,null)
    insert into temp_table values(8,'319','215','ABC','MNO','2015/01/01','2015/01/01','17:00:00','20:00:00',18,11,60,'21:00:00',null,null)
    --output should come as the data in temp_table_final
    create table temp_table_final
    row_id int,
    engine_no varchar(20),
    schedule_no varchar(20),
    start_station varchar(20),
    end_station varchar(20),
    startdate datetime,
    enddate datetime,
    starttime datetime,
    endtime datetime,
    record_id int,
    engine_id int,
    Mgt int,
    nextstarttime datetime,
    Schedule_ref varchar(20),
    Engine_Ref varchar(20)
    GO
    insert into temp_table_final values(1,'19O','101','SGC','IXP','2015/01/01','2015/01/01','00:00:00','01:00:00',1,10,60,'02:00:00','101','19O-1')
    insert into temp_table_final values(2,'19O','102','SGC','IXP','2015/01/01','2015/01/01','00:30:00','01:30:00',2,10,60,'02:30:00','102','19O-2')
    insert into temp_table_final values(3,'19O','103','SGC','IXP','2015/01/01','2015/01/01','02:30:00','03:30:00',3,10,60,'04:30:00','103','19O-3')
    insert into temp_table_final values(4,'19O','104','IXP','DFW','2015/01/01','2015/01/01','03:30:00','04:00:00',4,10,60,'05:00:00','101','19O-1')
    insert into temp_table_final values(5,'19O','105','IXP','DFW','2015/01/01','2015/01/01','04:30:00','05:30:00',5,10,60,'06:30:00','102','19O-2')
    insert into temp_table_final values(6,'19O','106','DFW','ABC','2015/01/01','2015/01/01','05:00:00','06:00:00',6,10,60,'07:00:00','104','19O-1')
    insert into temp_table_final values(7,'19O','107','DFW','ABC','2015/01/01','2015/01/01','06:00:00','07:00:00',7,10,60,'08:00:00','107','19O-4')
    insert into temp_table_final values(8,'19O','108','DFW','ABC','2015/01/01','2015/01/01','07:00:00','08:00:00',8,10,60,'09:00:00','105','19O-2')
    insert into temp_table_final values(9,'19O','109','ABC','DEF','2015/01/01','2015/01/01','10:00:00','11:30:00',9,10,60,'12:30:00','106','19O-1')
    insert into temp_table_final values(10,'19O','110','XYZ','BDW','2015/01/01','2015/01/01','13:00:00','15:00:00',10,10,60,'16:00:00','110','19O-5')
    insert into temp_table_final values(1,'319','111','PQR','STU','2015/01/01','2015/01/01','00:00:00','01:00:00',11,11,60,'02:00:00','111','319-1')
    insert into temp_table_final values(2,'319','211','PQR','STU','2015/01/01','2015/01/01','04:30:00','15:30:00',12,11,60,'16:30:00','211','319-2')
    insert into temp_table_final values(3,'319','112','STU','ABC','2015/01/01','2015/01/01','05:00:00','06:00:00',13,11,60,'07:00:00','111','319-1')
    insert into temp_table_final values(4,'319','212','STU','DEF','2015/01/01','2015/01/01','06:00:00','07:00:00',14,11,60,'08:00:00','212','319-3')
    insert into temp_table_final values(5,'319','213','STU','PQR','2015/01/01','2015/01/01','07:00:00','08:00:00',15,11,60,'09:00:00','213','319-4')
    insert into temp_table_final values(6,'319','118','STU','XYZ','2015/01/01','2015/01/01','10:00:00','11:30:00',16,11,60,'12:30:00','118','319-5')
    insert into temp_table_final values(7,'319','119','DEF','JKL','2015/01/01','2015/01/01','13:00:00','15:00:00',17,11,60,'16:00:00','212','319-3')
    insert into temp_table_final values(8,'319','215','ABC','MNO','2015/01/01','2015/01/01','17:00:00','20:00:00',18,11,60,'21:00:00','112','319-1')
    insert into temp_table_final values(1,'19O','101','SGC','IXP','2015/01/01','2015/01/01','00:00:00','01:00:00',1,10,60,'02:00:00','101','19O-1')
    insert into temp_table_final values(2,'19O','102','SGC','IXP','2015/01/01','2015/01/01','00:30:00','01:30:00',2,10,60,'02:30:00','102','19O-2')
    insert into temp_table_final values(3,'19O','103','SGC','IXP','2015/01/01','2015/01/01','02:30:00','03:30:00',3,10,60,'04:30:00','103','19O-3')
    insert into temp_table_final values(4,'19O','104','IXP','DFW','2015/01/01','2015/01/01','03:30:00','04:00:00',4,10,60,'05:00:00','101','19O-1')
    insert into temp_table_final values(5,'19O','105','IXP','DFW','2015/01/01','2015/01/01','04:30:00','05:30:00',5,10,60,'06:30:00','102','19O-2')
    insert into temp_table_final values(6,'19O','106','DFW','ABC','2015/01/01','2015/01/01','05:00:00','06:00:00',6,10,60,'07:00:00','104','19O-1')
    insert into temp_table_final values(8,'19O','108','DFW','ABC','2015/01/01','2015/01/01','07:00:00','08:00:00',8,10,60,'09:00:00','105','19O-2')
    insert into temp_table_final values(9,'19O','109','ABC','DEF','2015/01/01','2015/01/01','10:00:00','11:30:00',9,10,60,'12:30:00','106','19O-1')
    insert into temp_table_final values(1,'319','111','PQR','STU','2015/01/01','2015/01/01','00:00:00','01:00:00',11,11,60,'02:00:00','111','319-1')
    insert into temp_table_final values(2,'319','211','PQR','STU','2015/01/01','2015/01/01','04:30:00','15:30:00',12,11,60,'16:30:00','211','319-2')
    insert into temp_table_final values(3,'319','112','STU','ABC','2015/01/01','2015/01/01','05:00:00','06:00:00',13,11,60,'07:00:00','111','319-1')
    insert into temp_table_final values(4,'319','212','STU','DEF','2015/01/01','2015/01/01','06:00:00','07:00:00',14,11,60,'08:00:00','212','319-3')
    insert into temp_table_final values(5,'319','213','STU','PQR','2015/01/01','2015/01/01','07:00:00','08:00:00',15,11,60,'09:00:00','213','319-4')
    insert into temp_table_final values(6,'319','118','STU','XYZ','2015/01/01','2015/01/01','10:00:00','11:30:00',16,11,60,'12:30:00','118','319-5')
    insert into temp_table_final values(7,'319','119','DEF','JKL','2015/01/01','2015/01/01','13:00:00','15:00:00',17,11,60,'16:00:00','212','319-3')
    insert into temp_table_final values(8,'319','215','ABC','MNO','2015/01/01','2015/01/01','17:00:00','20:00:00',18,11,60,'21:00:00','112','319-1')
    What we are doing here is generating a schedule for Trains departures.
    here, we should identify the train schedules making a chain of stations considering the endstation of a train engine no should be startstation of another record for same engineno. also the starttime of engineno should be nearest of nextstarttime
    of same station. 
    for example : if we pick Ist row "SGC-IXP", nextstarttime for same is "02:00:00 am". this means train departed from SGC will reach to IXP and is available for departure from IXP after "02:00:00 am". So we have to consider
    the record having startstation as IXP and nearest starttime to nextstarttime ("02:00:00"). So the next train departure would be IXP-DFW having starttime as "03:30:00 am".
    here you can see we have to assign the scheduleno of previously considered record to the chained schedule so we have updated the "schedule_ref" as 101. Also we have to assign the engine no - <counter of integer> to a single chain on schedule
    given in engine_ref.
    Regards
    Kumud

  • Use condition for assigning query to data provider with Javascript

    Hi gurus,
    I want to assign a query dynamically to a data provider when the user activated a Tab Panel in a Web Template (BI7).
    I can do that easily with the standard function SET_DATA_PROVIDER_PARAMETERS but this action is do each time of the activation of the Tab Panel.
    So I try to make a javascript for assign the query to the data provider only one time.
    My problem is I havenu2019t found a solution to get the default query assigning to the data provider.
    Itu2019s an example of what I try to do with javascipt :
    function Load_Query()
    var r=GET_DATA_PROVIDER PARAMETER ("QUERYVIEW_DATA_PROVIDER")
    if (r <>empty)
        executeJS_SET_DATA_PROVIDER_PARAMETERS_R()
    Thank you for help,
    Franck

    Hi Janice,
    You can check the Option Display Variable values only once.
    Thie you will the variables being displayed only once eventhough they are used in Multiple Data provider.
    If you want som variable to be data provider specific, then , you can create a new variable and add it int eh query.
    For example in your case let us say COMP1 is the variable for company code used in DP1 and DP2.
    If you want a different values to be selected for company code in DP1 and DP2 just replce the variable in DP2 with a new variabel COMP2.
    Hope this helps.
    Regards.
    Shafi.

Maybe you are looking for