OEM Job schedule definition to Flat file

Can anyone advice me on the best method of extracting a OEM Job schedule definition - from an existing OEM server - to a flat file executable deployment script i.e. SQL*Plus script, schell script e.t.c.
Your help will be most appriciated.

Thanks for your response. Is this fuctionality usable after OEM Job creation?
Message was edited by:
kayvic01

Similar Messages

  • How to upload  schedule line from flat files to sap file

    dear all,
    i want to upload the schedule lines from flat files to sap schedulle lines
    but the flat files have 15 schedule lines and the data is as per date
    so how to upload that and the fields available in flat files are more than the sap screen
    we are having more than 6 items
    and 15scedule lines its abt 90data to be upload
    for one customer in every 15 day
    so how to do this
    is there any direct use in functional side
    with out the help of any abap
    but my user will do it
    so he need a permanent solution
    with regards
    subrat

    Hi Subrat ,
    u can upload the data either ( Master /Transaction) data with the help of lsmw. for that all u need to do is go through the lsmw and do it. in that u can go Batch input recording/ BAPI/ IDOC any of that. here i am sending the LSMW Notes go through it and do the work.
    once u create the LSMW project then u can ask the data from user or u can explain the user about the program and can run the flat file to upload the data.
    if u require LSMW material Just send me blank mail from u. my mail id is [email protected]
    Reward if Helpful.
    Regards,
    Praveen Kumar.D

  • SAP BO DI 3.2 How to import definition for flat files

    Hi All,
    I have several MS Access database with their respective tables , but Im needing now to import the tables structure/definitions of their tables as Flat Files definitions inside SAP BO DI 3.2 , is there an automatically way to do this without using the Flat File Editor?
    Thanks in advance!

    Hi,
    if you have the data in an Access database can you not just connect via ODBC to get the data?  This would also allow you to get the column metadata easily.
    If you are going to export the data from Access to flat files then do this first and then use the Flat File editor to create each file in turn, as long as you have access to the file then you can open it in the file editor and Designer will import the schema for you.
    regards,
    Adrian

  • OEM Scheduled Job Not Assuming OS User File Permissions

    I have defined and scheduled a job using OEM Grid Control that executes an OS level command which is a script containing the HP-UX find command. When I run the script at the OS level the script completes successfully. When the job scheduler runs the job it fails because the find command doesn't have permission to recursively traverse the subdirectories even though the subdirectories are owned by the process executing the find command. The preferred credentials for the job are set to the same user that I used to run the job manually. Why would this be happening?
    We're using OEM Grid Control 10.2.0.3.0 running on Linux used to manage Oracle 10G on HP-UX. Please help.

    Please paste the script and error

  • Load data from flat file to target table with scheduling in loader

    Hi All,
    I have requirement as follows.
    I need to load the data to the target table on every Saturday. My source file consists of dataof several sates.For every week i have to load one particular state data to target table.
    If first week I loaded AP data, then second week on Saturday karnatak, etc.
    can u plz provide code also how can i schedule the data load with every saturday  with different state column values automatically.
    Thanks & Regards,
    Sekhar

    The best solution would be:
    get the flat file to the Database server
    define an External Table pointing to that flat file
    insert into destination table(s) as select * from external_table
    Loading only single state data each Saturday might mean troubles, but assuming there are valid reasons to do so, you could:
    create a job with a p_state parameter executing insert into destination table(s) as select * from external_table where state = p_state
    create a Scheduler chain where each member runs on next Saturday executing the same job with a different p_stateparameter
    Managing Tables
    Oracle Scheduler Concepts
    Regards
    Etbin

  • Background job finished but flat file not created in the Background

    Dear all,
    ZHR_CSD program is scheduled to run daily at 00:01:00. This program is generating the flat file in the folder CSD/HR.
    when i  schedule this program to run immediately it is generating the flat file.
    But when i schedule this program in Background it is not generating flat file .
    Regards

    Hi,
    As suggested by Eric, your Z Program is probably using GUI_DOWNLOAD Function Module.
    GUI_DOWNLOAD or any other GUI function modules (FM) will only run in foreground, not in Background.
    Ask your developer to code that Z Program with OPEN DATASET logic, if its possible.
    The Reverse situation is well described in [this thread|Background Job assigment with variant in SM37 , for Textfile uploading], please refer it to get some relative information for the same.
    Regards,
    Bhavik G. Shroff

  • Extract data from Query to a Flat File - scheduling in the background

    Hi All,
    We need to extract data from a query to a flat file periodically, but each time 10 separate executions based on 10 profit centers, to get the separate files.
    So, we have PC1, PC2, ... PC10 Profit centeres.
    We have Query1, with profit center variable.
    We have created 10 variants for Query1 so that we can execute for 10 profit centers.
    Now, we have used RSCRM_BAPI to execute the Query1 and write to a flat file. Now we are having a problem in getting this feature used to execute the query automatically for 10 different varaints.
    Thanks in advance,
    _ Shashi

    Shashi,
       You can try using Information Broadcasting or Reporting Agent. You can schedule this in Information Broadcasting.
    check this link for Information Broadcasting... <a href="http://help.sap.com/saphelp_nw04s/helpdata/en/a5/359840dfa5a160e10000000a1550b0/content.htm">http://help.sap.com/saphelp_nw04s/helpdata/en/a5/359840dfa5a160e10000000a1550b0/content.htm</a>
    Nagesh Ganisetti.

  • Regarding Background Job scheduling for file to file scenario

    Hi Guru's
    Good morining all of u
    I have one doubt on ...
    can we do background job scheduling for file to file scenario?
    Please give me response as early as possible.
    thanks and regards
    sai

    background jobscheduling for file adapter...
    you can schedule the file adapter according to your requirement as follows:
    In the  Communication Channel Monitoring Locate the link Availability Time Planning,  In Availability Time Planning, choose the Availability time as daily and say create, give the time details, select the communication channel in your case file adapter , goto the Communication Channels tab and filter and add the respective channel, save it
    /people/shabarish.vijayakumar/blog/2006/11/26/adapter-scheduling--hail-sp-19-

  • Scheduling Problem for uploading Data from Flat file to SAP

    Hi guys,
    I am facing a weared problem in uploading some leave records in z table. The code is working fine if we run it through se38 after selecting the file from a shared location from production server which has all the access rights.
    This folder lies in the \usr folder of SAP Production.
    I have kept all the Flat files in the shared path "
    Tis-mum-iz-s1\migration\SAP-INT\leave\" ...
    To give u exact directory structure..
    Tis-mum-iz-s1 is the Server Name
    usr is the SAP System folder used for uploads and downloads
    usr |
    ...-> Migration |
                      -> SAP-INT |
                                 -> leave -> (Flat Files)
    Migration folder is shared with all rights.
    Obviously, we cannot give shared drive as the variant in the scheduler.
    So i use the system path i.e. \usr\sap\tmp\migration\sap-int\leave\ as the variant.
    All my other download programs are working fine with this path as a variant...
    But my this particular upload program does not work with this path....
    I am giving u my code...
    TATA INTERACTIVE SYSTEMS (A Division of TATA INDUSTRIES LIMITED)
    REPORT      :  ZMIGRATE_ZLEAVE
    DESCRIPTION :  To Upload the Leave data. (ZLEAVE)
    CREATED BY  :  Abhishek Bachhawat
    CREATED ON  :  01.09.2005
    CONSULTANT  :  ANAND
    REPORT  ZMIGRATE_ZLEAVE.
    TABLES: ZLEAVE.
    data: begin of wtab,
              MANDT(3),
              ZLVID(8),
              PERNR(8),
              ZSTDT(8),
              ZENDT(8),
              ZDAYS(4),
              AEDAT(8),
              ERDAT(8),
          end of wtab,
          itab like WTAB occurs 0 WITH HEADER LINE.
    data: temp like zleave occurs 0 WITH HEADER LINE.
    SELECTION-SCREEN BEGIN OF BLOCK file
                   WITH FRAME TITLE text-005.
    parameters: file like rlgrap-filename Obligatory.
    Concatenate File SY-DATUM '_Leave.txt' into File.
    SELECTION-SCREEN END OF BLOCK file.
    at SELECTION-SCREEN ON VALUE-REQUEST FOR file .
      CALL FUNCTION 'WS_FILENAME_GET'
        IMPORTING
          FILENAME = file.
      IF SY-SUBRC <> 0.
      ENDIF.
    start-of-selection.
      if file ne space.
        CALL FUNCTION 'WS_UPLOAD'
          EXPORTING
            FILENAME = FILE
            FILETYPE = 'DAT'
          TABLES
            DATA_TAB = ITAB.
      else.
        message e000(zps) with 'Specify a file'.
      endif.
      SORT ITAB BY ZLVID.
      LOOP AT ITAB.
        REFRESH TEMP.
        CLEAR TEMP.
        TEMP-MANDT = sy-mandt.
        TEMP-ERDAT = SY-DATUM.
        TEMP-ZLVID = ITAB-ZLVID.
        TEMP-PERNR = ITAB-PERNR.
        TEMP-ZSTDT = ITAB-ZSTDT.
        TEMP-ZENDT = ITAB-ZENDT.
        TEMP-ZDAYS = ITAB-ZDAYS.
        TEMP-AEDAT = ITAB-AEDAT.
        TEMP-ERDAT = ITAB-ERDAT.
        APPEND TEMP.
        SELECT SINGLE *
               FROM   ZLEAVE
               WHERE  ZLVID = TEMP-ZLVID
               AND    PERNR = TEMP-PERNR.
        IF SY-SUBRC = 0.
          UPDATE ZLEAVE SET ZSTDT = TEMP-ZSTDT
                            ZENDT = TEMP-ZENDT
                            ZDAYS = TEMP-ZDAYS
                            AEDAT = TEMP-AEDAT
                            ERDAT = TEMP-ERDAT
                 WHERE ZLVID = TEMP-ZLVID
                 AND   PERNR = TEMP-PERNR.
        ELSE.
          INSERT ZLEAVE FROM TABLE TEMP.
          COMMIT WORK.
        ENDIF.
      ENDLOOP.

    Hi,
    open dataset file for input in text mode.
    check sy-subrc = 0.
    while sy-subrc = 0.
      read dataset file into wa.
      if sy-subrc = 0.
      append wa to itab.
      else.
        exit.
      endif.
    endwhile.
    close dataset file.
    regards
    Siggi
    PS: check also the F1-help for open, read and close statements!

  • SSIS package works in development environment but fails when job scheduler executes, file path invalid

    SIS package works in development environment but fails when job scheduler executes, file path invalid
    Relatively simple package to get OLE-DB connection to MS FoxPro 9.0 DB
    The failure in the job log states that the path is invalid its a network path (\\192.168.1.xxx\foldername) this has been run several ways with the remote computer being mapped as a network drive and through the \\ notation described above.
    Thinking it was a security issue between the SQL agent account and my account I tested by subsitution myself as a proxy account for the agent when running this job, again same result failed on network path.
    One issue I see is that the remote computer is running Server 2000 (legacy software incompatable with newer versions) 
    Is it possible that this is a security issue, since if i understand correctly the current MS domain security model didn't exist until Server 2003.

    Hi REIData,
    Have you got the issue resolved? Based on your description, please make sure the target folder is shared properly. If the computer on which the SQL Server Agent job runs doesn’t join the domain as the server that hosts the shared folder, you have to share
    the folder with everyone by adding “Everyone” to the people list on the File Sharing page of the folder and assign "Read/Write" permission to it.
    Regards,
    Mike Yin
    TechNet Community Support

  • Best SAP Security Practices Print,file,job schedule, archiving

    Hello All, i would like to know in your experience which will be the best practices for Security  for this list below:
    - Printer security (especially check printing)
    - File path security for export/import
    - Best Practice for Job Schedule and Spool file
    - Archiving process (I can't think of any specific to security, other than Security Audit Logs)
    Are there any special transactions/system settings/parameters that must be on place in order to hard SAP Systems?
    Do you have any documentation related?
    I mean for example Job, spool i think user must just only run heir own jobs,and se their own works for printing, is there a paremeter to athenticate Prints/user, etc.
    Please let me know your comments about those related issues.
    I appreciate your help.
    Thanks a lot.
    Ahmed

    Hi,
    PFCG_TIME_DEPENDENCY
    This is best to run once a day mostly after 12.01 am as it removes the roles which are invalid for current date. As role assignment is on date basis there is no advantage of running it hourly.
    /VIRSA/ZVFATBAK
    This is for GRC 5.3, and this job is to collect FFID logs from backend to GRC repository, so if you have frequent FFID usage you can schedule it hourly or for every 30 min too, if you have enough bandwidth in your server to get the latest log report. or else you can have it scheduled for twice a day too, so it is purely based on your need.
    Hope this helps.
    BR,
    Mangesh

  • OEM 11g The Notification Schedule; Edit Schedule Definition buttons does not work

    When I go to the Edit Schedule Definition in OEM 11g and check the Days of Week schedule and try to hit the Batch-Fill-in button or any of the other buttons, nothing happens. It's as if the buttons
    are not working. I could really use some help on this.

    Hi,
    Are you using EM 11g Dbcontrol
    Is yes then we have a known issue marked in BUG 13907733
    Refer to below doc and apply the recommended patch
    Dbcontrol Email Notifications Are not received for metric alerts with a valid Notification Rule (Doc ID 1927762.1)
    Regards,
    Rahul

  • Flat file Background Job.

    Hi,
    I want to load flat file data in SAP BI through background job. so what is  the process to save the flat file in the application server home directory(DIR_HOME) (trans-AL11). ???
    Regards,
    Viraj.

    Hi,
    See this link
    Flatfile loading into Application Server:-
    Uploading and Downloading Excel file from application server.
    Regards
    Tg

  • Futures on Scheduling Flat File Refreshes

    Using Rich Client to bring in flat file data.
    Is there any way to
    - refresh flat file from Infoview from shared drive or
    - schedule refresh of Rich Client report and post to Infoview?
    Any plans to add this as would like to 'schedule' the refreshing of a report that contains flat file data.

    Hi,
    It seems that the flat file and data source are not in sync.... check your flat file format and prepare it in sync with your data source.
    I guess you have an amount field followed by currency in your flat file, but in your data source currency field is not next to the amount KF or vice versa. It could also be that the KF you have defined is not of Amount type, and hence it doesn't expect the currency field.
    Hope this helps
    Godhuli

  • Question about flat file schema definitions

    Hello,
    I have defined a flat file schema which works fine. However, I got now a new requirement for this schema: It has to support future potential additional fields in the end of the records.
    The solution I used is quit "ugly". I added an additional filler at the end of the record and configured it as "minOccurs = 0" and set Allow early termination of optional fileds to true.
    This works but I don't like it.
    It seems to me that there must be a property for ignoring any additional fields after the last one, so I won't need  this filler field.
    Does anyone familiar with such option/ property?
    Thanks all.

    Hi David,
    Min occurs and allow_early_termination options for perfectly valid to mark the last record as optional. There is no issue with approach and its not ugly. I have seen this approach being implemented in many projects without any issues.
     Refer this MSDN blog on this context and what you're doing is correct.
    http://blogs.msdn.com/b/skaufman/archive/2004/05/07/127899.aspx
    If this answers your question please mark it accordingly. If this post is helpful, please vote as helpful by clicking the upward arrow mark next to my reply.

Maybe you are looking for