Reusing the job in odi 11.1.1.3

Hi All,
We have created an interface between a source table and target table.
Now with the same source table, we have the same target table(same coulmn names etc) in three different databses.
How can I reuse this interface/job.
Do I need to recreate the job again for three different databases.
or
I can use the same job by changing some connection details to the respective databases.
Thanks

Hi
It is posible with one interface to move data from 3 source DB to target DB for that you can apporch Multi context option to move data from multiple source DB to single target DB and as per your issue, you have same source table (same table name and same column) available in 3 source DB to extract it to single target table.
Create 3 Physical Architecture for 3 source DB in Topology Manager with same logical name in Logical Architecture and create 3 different context and assign 3 different context to 3 source DB in Physical Architeture.
While creating Logical Schema for target DB add 3 Logical schema as shown in below example
eg:
Source DB1:
Logical Schema: OracleDB
Context: Source1
Source DB2:
Logical Schema: OracleDB
Context: Source2
Source DB3:
Logical Schema: OracleDB
Context: Source1
Target DB1:
Logical Schema: TargetDB
Context: Source1
Context: Source2
Context: Source3
While executing interface, choosed context name for which source DB data will move to target DB
Note: Make source that for 3 source DB logical schema should same
Regards,
Phanikanth

Similar Messages

  • Sharing OS commands from the Job Library

    I am looking at setting up an OS command script in the Job Library. I want to maintain only one copy of the script. If I enter and save the script in the library, I can then use it when I go to submit a recurring job. However, it looks like when I submit the job based on the library it makes another copy of the script instead of maintaining a link to the version in the library. If I modify the library version, it looks like the recurring job based on the library job does not pick up the modified script. Is there a way to reuse the job library script, so only one version is maintained?

    Let me restate the question:
    I create a new OS command job in the library that searches for core files on the system. I save this new library job as Find_Core_Job.
    I now submit a job based on the library job, choosing a target, and scheduling it to run every night. A job is created called Find_Core_Job.1.
    After the job has been running for a few days I decide I want to make a change to it. I edit the library job and change the OS command and save the library job Find_Core_Job.
    Will the job I submitted, Find_Core_Job.1, pick up the change to the OS command that I made in the original library job or did it make a copy when I did the original submit. I only want to maintain one copy of the script in the library. I do not want to have to modify two copies of the OS command script.

  • Are the OPH webservices/odi migrated to  SOA FMW in the pip pmdm11.2?

    To integrate OPH (Oracle Product Hub 12.1.3) with RODOD using AIA Pre-Built Integrations Product Master Data Management 11.1 I installed an OAC (Oracle Application Server) where has been deployed the OPH webservice:
    PIMWebServices-ICC-WS
    PIMWebServices-Item-WS
    PIMWebServices-Valueset-WS
    which working with the ODI10 that is deplojed in the same server.
    My question is if for the new release 11.2 (about the integration of Oracle Product Hub 12.1.3 with Oracle Pre-Built Integrations Product Master Data Management 11.2) the webservices and ODI project have been migrated to the SOA FMW 11.1.1.6 platform or not?
    If not have I to install a new version of PIMWebServices and ODI project or I can reuse the same released for 11.1?
    Below the link of new Pre-Built Integrations 11.4 where I did not find the answer:
    http://docs.oracle.com/cd/E38316_01/index.htm

    Hi Ingrid,
    It's compatiable , but not in all cases. For example, If you are using business rule then such a deployment may give error.
    Sudiv

  • ** Is it possible to run a windows .bat script job from ODI between server?

    Hello Friends,
    We have ODI 10.1.3.5 on a Win2003 Server.
    Have a windows Batch script job .bat(this is a related to a different program - DRM) on a different Windows 2003 server...is it possible to execute .bat job thru ODI between servers?
    Appreciate your suggestions...
    Thanks.

    Hi AJ,
    Here are more details:
    Server-A: ODI Server (Win2003)
    Server-B: Oracle DRM Server (Win2003)
    Server-C: ODI scheduler agent is installed (Linux)
    We have a .bat batch script job on Server-B which we are planning to run from Server-A (as part of the ODI package execution flow)
    This .bat job exports the DRM hierarchy data to a table that ODI uses to process next steps.
    So, you are suggesting to install ODI agent on Server-B? If we do that...not sure if there be issues if we have to run the package with other scenario's (interfaces/Procedures) that refer agent isntalled in Server-C??
    Any feedbacks are appreciated!!
    Thanks

  • Need to run the JOB without Ineractive Screen

    Dear All,
    We have written a new ABAP program for executing the Macro background Job and used JOB_OPEN,JOB_SUBMIT and JOB_CLOSE Function Modules to run the JOB from ABAP Program.  We are able to run the JOB successfully for Any other ABAP Programs through the custom program, but when I tried with the same approach to run the Macro job through my new ABAP Program the JOB is throwing an error.
    Error Message is like below
    Job started
    Step 001 started (program /SAPAPO/TS_BATCH_JOB, variant TEST, user ID GKMN3511
    Specify a start time
    Job cancelled after system exception ERROR_MESSAGE
    Point of doubts:
    -- Can we really execute the Macro job through the ABAP program with out any dialogue screen (because when I use the "/SAPAPO/MC8G - Schedule Demand Planning in the Backgroundu201D program in my custom program it pop-ups a interactive screen which I don't want in my ABAP program while scheduling the macro job)
    Regards,
    Raju

    In transaction /sapapo/mc8t --> define the activity that looks at the macro in question
    In transaction /sapapo/ mc8d --> define the job that looks at that activity and a selection iD/aggregation level
    In transaction /sapapo/ts_batch_run --> save a variant pointing to the job defined in mc8d
    you can then run the /sapapo/ts_batch_run job with that variant

  • Check The Job Status

    Hi SAPsimhas,
    My requirement is as below.
    I will loop through an internal table and create some files with background processing method.(JOB_OPEN,SUBMIT & JOB_CLOSE)
    Now once this job is triggered I need to check the status for that job after maximum 10 mins. Its not neccessary that each job will take 10 mins. only,It can be finished in 2 mins also.
    Once the job is finished then I will XCOM that file.
    So my question is how to code such timer which will check the job status for that job? I have a FM BP_JOB_READ which returns the job status.
    Reply is highly appriciated.

    Try this way
      do.
        call function 'SHOW_JOBSTATE'
          exporting
            jobcount         = p_jobc
            jobname          = p_jobn
          importing
            aborted          = v_aborted
            finished         = v_finished
            ready            = v_ready
            running          = v_running
            scheduled        = v_scheduled
          exceptions
            jobcount_missing = 1
            jobname_missing  = 2
            job_notex        = 3
            others           = 4.
        if v_aborted eq 'X'.
          exit.
        endif.
        if v_finished eq 'X'.
          select single * from tbtcp into wa_tbtcp1
                     where jobname eq p_jobn
                       and jobcount eq p_jobc.
          if sy-subrc eq 0.
            move wa_tbtcp1-listident to p_spono.
          endif.
          exit.
        endif.
        if v_count le 10.
          call function 'ENQUE_SLEEP'
            exporting
              seconds = 60.
          v_count = v_count + 1.
        endif.
      enddo.

  • How to find out the user from the Jobs queue in Report server

    Hello All!
    I have a doubt about finding out the user from the scheduled jobs queue. Say I go ahead and schedule a report on Reports Server how can I find out the user name. When I view the jobs using showjobs I could see that the DBMS_JOBS table has a column under "Job Owner". But it invariantly shows it is "rwuser". So is there a way to find out which user has scheduled which job?
    Regards
    Shobha

    hi,
    The below tables will give only the name .
    USER_ADDRS
    USER_ADDR
    USER_ADDRP
    USR02
    i think you need email address .
    you can use this Tcode : su01d
    and give the user name and excute it
    i hope it will help you.
    Ram
    Edited by: Ram velanati on Jun 30, 2008 6:57 PM

  • BODS 3.1 : How to trigger an email alert for the jobs on BODS server ?

    Hi All.
    I have this request.
    BODS 3.1 : How to trigger an email alert for the jobs on BODS server ?
    We have jobs scheduled on BODS running smoothly and absolutely fine.
    But to check, i am logging into the admin console and check for the jobs status.
    I would like to have an email to be received from BODS after each job is finished.
    It could succuessful. Or it could fail.
    Whatsoever, i wish to receive an email alert as soon as a job is finished.
    Can anyone advise me as to whether this could be made possible.
    And if yes, how this could be done.
    Thanks for your help in advance.
    In BOE CMC / for webi / schedule / we find an option to send email for a job success or a job failure.
    Is there any option similar to that in BODS ?
    Also would like to know :
    how to use the smtp_to or mail_to functions ?
    how to set up the smtp server for this ?
    thanks
    REgards
    indu
    Edited by: Indumathy Narayanan on May 31, 2011 3:47 PM

    Hi.
    Since am new to this BODS. I need some help.
    I already have many jobs which are running absolutely fine.
    And when a job runs, and finishes, am able to see the trace saying
    e.g. :
    Job_abc is completed successfully.
    We got the smtp service activated for our test server.
    and we hae a group email id.
    I have put the details of the smtp server / ip address / and said apply restarted.
    The i created a simple test script as below :
    print (' Before email ' );
    smtp_to('abc@company_name.com', 'Job ' || job_name() ||' on ' || host_name() || ' has FAILED',
    ' the job has failed', 0, 0);
    print('After Email ');
    It does send a email to as per smtp_to whatever email is specified.
    But how to differentiate between a job success
    And a job which has failed.
    I wish to have a mail which says on the subject :
    'Job ' || job_name() ||' on ' || host_name() || ' has completed successfully'
    ==> IF it is a success
    OR
    'Job ' || job_name() ||' on ' || host_name() || ' has failed'
    ==> if it has failed
    How to make the system identify, whether
    to send a success message or a error message whatever
    Could anyone advise.
    thanks
    indu

  • No wifi connection on my 4s. Did a Reset however this didn't do the job as well

    ffor the last 24 hours my iPhone does not able to connect to available wifi networks.. I don't see any optional networks . I did a Reset however this didn't do the job as well.. Any sugestions ?

    Hi michal272,
    Welcome to the Apple Support Communities!
    I understand that your iPhone is not able to see any available Wi-Fi networks or connect to any Wi-Fi networks. I understand that you have already reset your iPhone and the issue continues. I would recommend continuing your troubleshooting by working through the steps in the attached article that may not have yet been completed, such as resetting your network settings. 
    If your iPhone, iPad, or iPod touch won’t connect to a Wi-Fi network - Apple Support
    Best regards,
    Joe

  • Job Cancelled with an error "Data does not match the job def: Job terminat"

    Dear Friends,
    The following job is with respect to an inbound interface that transfers data into SAP.
    The file mist.txt is picked from the /FI/in directory of the application server and is moved to the /FI/work directory of application server for processing. Once the program ends up without any error, the file is moved to /FI/archive directory.
    The below are the steps listed in job log, no spool is generated for this job and it ended up with an error "Data does not match the job definition; job terminated".Please see below for more info.
    1.Job   Started                                                                               
    2.Step 001 started (program Y_SAP_FI_POST, variant MIST, user ID K364646)                           
    3.File mist.txt copied from /data/sap/ARD/interface/FI/in/ to /data/sap/ARD/interface/FI/work/.
    4.File mist.txt deleted from /data/sap/ARD/interface/FI/in/.                                   
    5.File mist.txt read from /data/sap/ARD/interface/FI/work/.                                    
    6.PD-DKLY-Y_SAP_FI_POST: This job was started periodically or directly from SM36/SM37 (Message Class: BD, Message Number : 076)  
    7.Job PD-DKLY-Y_SAP_FI_POST: Data does not match the job definition; job terminated (Message Class : BD, Message No. 078)    
    8.Job cancelled after system exception
    ERROR_MESSAGE                                                
    Could you please analyse and come up about under what circumstance the above error is reported.
    As well I heard that because of the customization issues in T.Code BMV0, the above error has raised. 
    Also please note that we can define as well schedule jobs from the above transaction and the corresponding data is stored in the table TBICU
    My Trials
    1. Tested uplaoding an empty file
    2. Tested uploading with wrong data
    3. Tested uploading with improper data that has false file structue
    But failed to simulate the above scenario.
    Clarification Required
    Assume that I have defined a job using BMV0. Is that mandatory to use the same job in SM37/SM36 for scheduling?
    Is the above question valid?
    Edited by: dharmendra gali on Jan 28, 2008 6:06 AM

    Dear Friends,
    _Urgent : Please work on this ASAP _
    The following job is with respect to an inbound interface that transfers data into SAP.
    The file mist.txt is picked from the /FI/in directory of the application server and is moved to the /FI/work directory of application server for processing. Once the program ends up without any error, the file is moved to /FI/archive directory.
    The below are the steps listed in job log, no spool is generated for this job and it ended up with an error "Data does not match the job definition; job terminated".Please see below for more info.
    1.Job Started
    2.Step 001 started (program Y_SAP_FI_POST, variant MIST, user ID K364646)
    3.File mist.txt copied from /data/sap/ARD/interface/FI/in/ to /data/sap/ARD/interface/FI/work/.
    4.File mist.txt deleted from /data/sap/ARD/interface/FI/in/.
    5.File mist.txt read from /data/sap/ARD/interface/FI/work/.
    6.PD-DKLY-Y_SAP_FI_POST: This job was started periodically or directly from SM36/SM37 (Message Class: BD, Message Number : 076)
    7.Job PD-DKLY-Y_SAP_FI_POST: Data does not match the job definition; job terminated (Message Class : BD, Message No. 078)
    8.Job cancelled after system exception
    ERROR_MESSAGE
    Could you please analyse and come up about under what circumstance the above error is reported.
    As well I heard that because of the customization issues in T.Code BMV0, the above error has raised.
    Also please note that we can define as well schedule jobs from the above transaction and the corresponding data is stored in the table TBICU
    My Trials
    1. Tested uplaoding an empty file
    2. Tested uploading with wrong data
    3. Tested uploading with improper data that has false file structue
    But failed to simulate the above scenario.
    Clarification Required
    Assume that I have defined a job using BMV0. Is that mandatory to use the same job in SM37/SM36 for scheduling?
    Is the above question valid?

  • Destination disabled. []: [CrystalEnterprise.Ftp]. Please note the name of the job server used for your request and contact your system administrator to make sure the specified destination is enabled. (FWB 00031)

    Hi
    In BO 4.0 SP 9 when a administrator tries to schedule a report via CMC there is no error
    But when a user schedules a report and the destination is FTP location -> Use default settings he gets following error
    Destination disabled. []: [CrystalEnterprise.Ftp]. Please note the name of the job server used for your request and contact your system administrator to make sure the specified destination is enabled. (FWB 00031)
    There is only one Job Server and the destinations are enabled in it
    There is no Job server for Crystal Reports Job Server
    Do i need to create it and how.

    Please check if you have proper rights to schedule to FTP. You can create a new job server, whenever you schedule it, there are multiple job servers, it will handle based on the load. But it is not mandatory, depends on the load.

  • Service template problem - Unable to perform the job because one or more of the selected objects are locked by another job - ID 2606

    Hello,
    I’ve finally managed to deploy my first guest cluster with a shared VHDX using a service template. 
    So, I now want to try and update my service template.  However, whenever I try to do anything with it, in the services section, I receive the error:
    Unable to perform the job because one or more of the selected objects are locked by another job.  To find out which job is locking the object, in the jobs view, group by status, and find the running or cancelling job for the object.  ID 2606
    Well I tried that and there doesn’t seem to be a job locking the object.  Both the cluster nodes appear to be up and running, and I can’t see a problem with it at all.  I tried running the following query in SQL:
    SELECT * FROM [VirtualManagerDB].[dbo].[tbl_VMM_Lock] where TaskID='Task_GUID'
    but all this gives me is an error that says - conversion failed when converting from a character string to uniqueidentifier msg 8169, level 16, State 2, Line 1
    I'm no SQL expert as you can probably tell, but I'd prefer not to deploy another service template in case this issue occurs again.
    Can anyone help?

    No one else had this?

  • Unable to remove a host from VMM - Error (2606) Unable to perform the job because one or more of the selected objects are locked by another job.

    I am unable to remove a host from my Virtual Machine Manager 2012 R2. I receive the following error:
    Error (2606)
    Unable to perform the job because one or more of the selected objects are locked by another job.
    Recommended Action
    To find out which job is locking the object, in the Jobs view, group by Status, and find the running or canceling job for the object. When the job is complete, try again.
    I have already tried running the following command in SQL Server Management Studio
    SELECT * FROM [VirtualManagerDB].[dbo].[tbl_VMM_Lock] where TaskID='Task_GUID'
    I received this error back:
    Msg 8169, Level 16, State 2, Line 1
    Conversion failed when converting from a character string to uniqueidentifier.
    I have also tried rebooting both the host and the Virtual Machine Manager Server.  After rebooting them both, I still receive the same error when trying to remove the host.
    Here are my server details
    VMM Server OS = Windows 2012 Standard
    VMM Version = 2012 R2 3.2.7510.0
    Host OS = Windows 2012 R2 Datacenter
    Host Agent Version = 3.2.75.10.0
    SQL Server OS = Windows 2012 Datacenter
    SQL Version = 2012 SP 1 (11.0.3000.0)

    Hi there,
    How many hosts are you managing with your VMM server?
    The locking job might be the background host refresher job. Did you see any jobs in the jobs view, when the host removal job failed?
    If there is no active jobs in the jobs view when this host removal job fails, can you please turn on the VMM tracing, retry the host removal, and paste back the traces for the failed job (search for exception and paste the whole stack)?
    Thanks!
    Cheng

  • Print the job from the spool in abtch mode

    I want to be able to print a job from the spool in batch mode. In other words a job currently exists that creates a report that currently goes to a printer.
    The user wants the report to print on two different printers. This extra printer cannot be placed in the job, since SAP only allows one printer per output. What I would like is to find a SAP program that would accept a spool name and a printer destination. It necessary we can write a program that might execute a SAP function to do the same.
    SP01, the spool transaction, does not easily adapt itself to batch execution. I donu2019t want to create a bdc session using it.

    Hi,
    If the domain of these fields are of type SYTIME, then adding the fields like a numeric value should suffice.

  • How to open a URL without session ID and reuse the current browser session?

    Hi All,
    I have a question about HTMLDB 2.0
    How to open a URL without session ID and reuse the current browser session?
    That was the behaviour in HTMLDB 1.6 ...
    My usecase for this is the following:
    We have written an issue tracking application, which sends e-mail to the interested users, when something happens.
    In these email we've put a link to some page, with some parameters in the URL.
    The idea is for the user to be easy to click on the hyperlink and to see the details of the ticket.
    When the user clicks on such a link he is directed to a login screen (page 101) and he enters his Username and password, and is then forwarded to the details for the ticket.
    Then he receives another email (e.g. for another ticked). He clicks on the link and :
    a) in HTMLDB 1.6 he goes to the details as he didn't close his browser and session is remembered
    b) in HTMLDB 2.0 he is prompted to enter username, password with the username populated
    Please tell me how can I achieve the same behaviour in HTMLDB2.0 as it was in HTMLDB 1.6.
    I understand this change is somehow security related, althogh I don't understand how. If you can explain this either I would be very happy?
    Best regards,
    Mihail Daskalov

    Mihail - I detailed a couple of approaches here: Re: Application Link
    Scott

Maybe you are looking for

  • Evaluating OS X Lion Server

    Hi everyone, I am new to OS X Lion and plan to buy that soon. Few questions:- a) Will there be a mountain lion server version? should i wait for that? when is the release? b) I have few iMac, macbook pro mostly on lion. 1 macbook on snow leopard and

  • Brazil Notafiscal error at Display Invoice MIR4

    Hi Guys We have one problem with Brazil invoice MIRA  we are saving invoice ( We are entering NF type E1 in details tab) Later on MIR6 post the invoice documnet Once MIR4 dispaly invoice if you click notafiscal button Error saying that notafiscal doc

  • Choosing a pdf version

    How do I convert a Word doc into an earlier version of pdf - eg 1.3?    

  • Time Machine 1.1 sometimes deletes backups that should keep

    About two years ago I was intrigued about the disparition of several Time Machine backups. After changing to a larger external backup disk, I've been trying to follow the behavior of Time Machine. Overall, it seems to be quite predictable, correspond

  • Need CRM overview for SAP BI

    Hi, Shortly I am going to work on CRM Server. which I need to pull the data from SAP CRM System to BI System. I am new to CRM. Can any one explain me the overview of CRM like what are the datasoources, data targets and any relevant info. Thanks VSM