SM37 - "pause" V3-update job while importing data source and extract struct

Hi
Insted of deleting an hourly schedulated V3-update job and the recreating it (after a import of a extract structure and a data source) I came to think about whether it is possible to, kind of, pause the job that I otherwise would delete and recreate.
Is this possible some how?
My thought:
1. Find the job in released status.
2. Select the job and change it (top menu Job-->change)
3. Change the scheduling to some day in the future
4. Change job status from released to scheduled
5. Import the extract structure and data source.
6. Change the schedule job back to hourly run.
7. Now the job should run on hourly basis again....
By the way: what do I import first, the data source or the extract structure?
Thanks in advance and kind regards,
Torben

We did delete and recreate...

Similar Messages

  • Data source and Extract Structure

    Hi all,
    I have a doubt on Data source and Extract Str,
    when i used one info cube with some char and kf's
    after i did extraction shall we change the extract strcuture prequently other wise better to use all the predefined extract strcuture hide in data source.
    if i hide in datasource after extraction i want to use some of hide fields, so we can use it? if we can use how can we extract data for that specified hide fields to bw side.

    If you wont to create the genaric data source at that time
    you can create cube with some char and kf.
    after extraction you wont to modify the structure ? if you mofify
    the structure then replicate the data source and delete the data
    and upload the data from r/3 to bw.its not correct way every time modifying the
    structure.if you put hide means that fields are not coming to
    bw side.if put modify the hide field ,first goto rsa6 edit data source
    remove the hide check box and save after come to bw replicate the data source
    delete the data and load the data (if you wont full data for hide field)
    if you dont load full load means only available upto data only what ever field you modify the
    hide to unhide,

  • Probelms while replicating Data Source and transport into test system

    I find  the following error messages whilre transporting into test system.
    DataSource 0EC_PCA_1 does not exist in source system of version A.
    Mapping between data source 0EC_PCA_1 and source system  is inconsistent.
    DataSource 0EC_PCA_1 does not exist in source system  of version A.
    Field AFABE will not be delivered from DataSource 0EC_PCA_1 in source system.
    any ideas please..........
    Enhanced the Extract structure and activated the data source in Source system
    Replicate the Data source in BI  and install the Infosource through BI Content  activated and transport them to Test system.

    Hi,
    1.First transport the data source from R/3 dev to R/3 quality
    2.Replicate the data source in BI Quality ( make sure you that in the data source tab, you are searching the data source below the appropriate source system ---otherwise you  data source may not appear)
    3.goto rsd1in BI development for activataing the info objects which you have added extra in the info source (z info objects )
      activate these info objects and collect in another request
    4.transport these info objects from BI dev to  BI quality
    5.transport entire data source to bw quality.
    Thanks,
    Tarun Brijwani.
    Edited by: Tarun Brijwani on Apr 6, 2009 4:26 PM
    Edited by: Tarun Brijwani on Apr 6, 2009 4:26 PM

  • SSIS 2012 is intermittently failing with below "Invalid date format" while importing data from a source table into a Destination table with same exact schema.

    We migrated Packages from SSIS 2008 to 2012. The Package is working fine in all the environments except in one of our environment.
    SSIS 2012 is intermittently failing with below error while importing data from a source table into a Destination table with same exact schema.
    Error: 2014-01-28 15:52:05.19
       Code: 0x80004005
       Source: xxxxxxxx SSIS.Pipeline
       Description: Unspecified error
    End Error
    Error: 2014-01-28 15:52:05.19
       Code: 0xC0202009
       Source: Process xxxxxx Load TableName [48]
       Description: SSIS Error Code DTS_E_OLEDBERROR.  An OLE DB error has occurred. Error code: 0x80004005.
    An OLE DB record is available.  Source: "Microsoft SQL Server Native Client 11.0"  Hresult: 0x80004005  Description: "Invalid date format".
    End Error
    Error: 2014-01-28 15:52:05.19
       Code: 0xC020901C
       Source: Process xxxxxxxx Load TableName [48]
       Description: There was an error with Load TableName.Inputs[OLE DB Destination Input].Columns[Updated] on Load TableName.Inputs[OLE DB Destination Input]. The column status returned was: "Conversion failed because the data value overflowed
    the specified type.".
    End Error
    But when we reorder the column in "Updated" in Destination table, the package is importing data successfully.
    This looks like bug to me, Any suggestion?

    Hi Mohideen,
    Based on my research, the issue might be related to one of the following factors:
    Memory pressure. Check there is a memory challenge when the issue occurs. In addition, if the package runs in 32-bit runtime on the specific server, use the 64-bit runtime instead.
    A known issue with SQL Native Client. As a workaround, use .NET data provider instead of SNAC.
    Hope this helps.
    Regards,
    Mike Yin
    If you have any feedback on our support, please click
    here
    Mike Yin
    TechNet Community Support

  • Error while importing data from XML file to a Oracle database

    I am getting the following error while importing data
    *** atg.xml.XMLFileException: No XML files were found for "/Dynamusic/config/dynamusic/songs-data.xml". 
    The files must be located  under the name "/Dynamusic/config/dyna
    *** java.io.FileNotFoundException: D:\MyApp\ATG\ATG10.0.3 (Access is denied)
    java.io.FileNotFoundException: D:\MyApp\ATG\ATG10.0.3 (Access is denied)
            at java.io.FileInputStream.open(Native Method)
            at java.io.FileInputStream.<init>(FileInputStream.java:120)
            at atg.adapter.gsa.xml.TemplateParser.importFiles(TemplateParser.java:6675)
            at atg.adapter.gsa.xml.TemplateParser.runParser(TemplateParser.java:5795)
            at atg.adapter.gsa.xml.TemplateParser.main(TemplateParser.java:5241)
    I have placed the FakeXADataSource and JTDataSource properties files as required. In fact I have been able to import data in all machines. Just this one machine is giving problems. I have also checked out the access rights. Can someone help me please?

    Confirm that you have access to D:\MyApp\ATG\ATG10.0.3 and existence of Dynamusic/config/dynamusic/songs-data.xm

  • Logon Error while importing data

    Hello,
    I have "Countries" flat table in my repository containing 2 fields Name & ISO Code.
    I wanted to import data i.e. Name & ISO Code in this flat table via Import Manager.
    I have excel file containing Name & ISO Code data.
    While importing data, I am getting an error:- "Logon Error: External table is not in the expected format."
    Would anybody know about this issue?
    Regards,
    Vikas

    Hi Vikas,
    Just check whether yr file .xls or .xlsx.
    If xlsx then change it to xls and check if problem is solved .
    If not try with disable macros.
    Please revert with yr results.
    Thanks,
    sudhanshu

  • While import data through ff_5 i am getting error

    hello
    while import data through ff_5 i am getting error  as below
    message error fv150 and fv151
    'Termination in statement no. 00009 of acct
    1101200570116; closing record 62F missing'
    so please give solution
    thank inadvance
    SIRI

    Dear Siri,
    I guess you are importing an MT940 format. This format should have the closing balance at the end.
    This closing balance starts with :62F:
    A sample is
    :61:0801180118DR3835,97NTRF000000//000000//
    :86:1022  LTD CHENNAI18012008
    :61:0801180118DR69885,09NCHK850819//850819//
    :86:6101  LTD COCHIN18012008
    :62F:C080118INR7210048,86
    I guess that is missing in the import file.
    Maintain that and the import will happen.
    Assign points if useful
    regards
    Venkatesh

  • Problem while importing data schema not found

    Hi,
    I'm facing problem while importing data through system to a schema and impdp gives me error that schema XYZ was not found but when I query dba_users schema exist there what could be the reason of this.
    Regards,
    Abbasi

    Hi,
    Here is complete listing
    bash-3.2$ impdp system/abc schemas=XYZ DIRECTORY=data_dir DUMPFILE=Today_dump.dmp LOGFILE=Today_log.log PARALLEL=8
    Import: Release 11.1.0.7.0 - 64bit Production on Friday, 07 January, 2011 11:18:57
    Copyright (c) 2003, 2007, Oracle. All rights reserved.
    Connected to: Oracle Database 11g Enterprise Edition Release 11.1.0.7.0 - 64bit Production
    With the Partitioning, OLAP, Data Mining and Real Application Testing options
    ORA-39002: invalid operation
    ORA-39165: Schema XYZ was not found.
    Regards,
    Abbasi

  • Trunccolumn value while importing data to physical layer in RPD

    Hi All,
    While importing data from database, I want to trunc few column values so that I can put complex join, but I am not getting any link to do so.
    Can anyone suggest me that if on database, if two tables are linked by outer join, than how can I restore the same join state in Physical layer while maintaining star schema.
    Thanks in advance.

    Hi,
    I am using version 10.1.3.4.0
    This query is a select type physical table in a physical layer of the RPD
    select distinct project_no,project_name from cw_vertical_d where project_no = VALUEOF(NQ_SESSION."Project_Number")
    NQ_SESSION."Project_Number" is a session non system variable.I am trying to set this variable through a dashboard prompt ( as a request variable ).
    So when a user selects any value in the dashboard prompt then this session non system variable "Project_Number" should be set and then BI Server should fire a query with this value to pick the data from database.
    But this report is failing with "*[nQSError: 23006] The session variable, NQ_SESSION."Project_Number", has no value definition*" error.
    Don't know what is causing this?
    Regards,
    Vikas

  • Error while activating data source in BI

    Hello Team
                     I am facing issues while activating data source. I followed following steps .
    a)First i created my target
    b)Then i made my Primary key as info provider
    c)Then i created one source system under file.
    d)Then i went to data source and created one application component
    e)In this application component i created my data source
    f)In this data source i imported my file and all the settings. Now when i am checking then it is not giving any error . But when i am trying to activate it it is giving error . Please find the supporting screesnhots . Can anyone please suggest me as to what is going wrong .

    Hi,
    Check the gateway settings for
    Gateway Host
    Gateway service
    Also check the UDC connection information.
    Check these links:
    1 - http://help.sap.com/saphelp_nw04/helpdata/en/78/ef1441a509064abee6ffd6f38278fd/content.htm
    2 - http://help.sap.com/saphelp_srm40/helpdata/es/54/ddfb3ff778ed28e10000000a1550b0/content.htm
    3 - https://encrypted.google.com/url?sa=t&source=web&cd=7&ved=0CCwQFjAG&url=http%3A%2F%2Fwww.bi-  
    expertonline.com%2Fdownloads%2FSchroeder0604.doc&rct=j&q=ud%20connection%20in%20sap%20bi&ei=WMDnTdm_E4ruvQOEzZWcDg&usg=AFQjCNGvqKuoTq9kypNMNGhis8WrZ-gFCg&cad=rja
    Let us know in case the issue still persists.
    Debanshu

  • Error while activating Data source

    Hi,
    I hit an error while activating data source with adapter 'WEB_PUSH web service'.
    The error message is 'Generation of conversion exists and PSA update ...' with details 'Error generating Web service /BIC/CZZZ_0010000100 ...' and 'Error when activating DataSource ZZZ_001'.
    I can't find much detail on this error.
    Has anyone hit this error before and if you did can you please let me know what to check?
    Thks & Regards,
    Bbee

    Hi,
    I am having the same problem.  Have you been able to solve the issue since you posted this last year? 
    Thanks,
    TCT

  • System Throws short dump while executing data source 2LIS_11_VASCL in RSA3.

    Hello Experts,
    I got short dump while executing data source 2LIS_11_VASCL in RSA3.
    The details are as follow.
    When importing object "MC11VA0SCL_TAB", the structure did not match the
    structure of the target object. The error occurred in component
    no. 567.
    This may be due to one of the following reasons:
    - the structure is actually different (the next field has, for
      example, a different data type) or
    - the same sequence of components were exported and imported,
      but the fields were from other include units.
    I have checked source and target fields and those are of same length and data type.
    Could you please give some points why this short dump occurs.
    Thanks
    Konda Reddy

    Hi,
    This error is not between BI and R/3.  It is within R/3.  You may have setup data and then modified the datasource.
    Delete setup data, generate your datasource again.  Setup data and try again.
    If you are on a live system with data setup after a downtime and you do not want to delete it, then bring your datasource to the original state when you setup data and pull data.
    Cheers....

  • Problem with delta update on customised export data source.

    Hi all BW gurus,
    I have created several customised export data sources in the R/3 system. And I have made use of the utility program <Z_CHANGE_DELTA_PROCESS> to update the ROOSOURCE table.
    While this can result in being able to do initial update (instead of full update), the subsuquent delta update is not working.
    In RSA7, the datasource is with green light and in RSA3, the extractor can extractor data out of it.
    Does anybody have some idea on this?
    I heard that there is a way to change the Business Transaction Event (BTE), but I don't quite understand how this is performed. And wonder if this is the only way out. To me, the triggering point would be the same, meaning the InfoPackage would be executed, then call the extractors in R/3 to extract the delta update in order to load into the InfoCube.
    And how do others do for executing delta update of the customised data source?
    Thanks a lot!!
    The utility program of <Z_CHANGE_DELTA_PROCESS>:
    =================================================
    report z_change_delta_process .
    *P_DATAS DataSource
    *P_DELTAP Delta Process for DataSource
    parameters:
      p_datas type roosource-oltpsource,
      p_deltap type roosource-delta.
    tables:
      roosource.
    data:
      ls_roosource type roosource.
      if p_datas(2) ne 'Z_'.
       message 'The DataSource needs to begin with ''Z_''.'  type 'E'.
      endif.
    select single * from roosource into ls_roosource
           where oltpsource = p_datas
           and objvers = 'A'.
    if sy-subrc eq 0.
       ls_roosource-delta = p_deltap.
       update roosource from ls_roosource.
    message 'The DataSource has been updated successfully.' type 'I'.
    else.
       message 'The DataSource entered is not valid, try again.' type 'E'.
    endif.

    Doesn't anyone have any idea on this topic?
    Appreciate for all the help. Thanks.

  • Not enough memory for Data Provider-Error while creating Data Source

    Hi,
    I am loading data into Master Data_Attribute InfoObject I am getting following error message while creating Data Source under "Proposal" Tab
    "Not enough memory for Data Provider"
    My Master Data InfoObject having 65 attributes
    My CSV file having 15,00000 records
    I am using BI 7.0 version
    If anybody faced this problem. Please share with me
    Thanks.

    Hi
    Here the problem with the space so plz contact ur BASIS people to increase the spae for particular object.

  • Time out while activating Data Source

    Hi All,
    I am faced with a problem. I am trying to load a cube from a flat file Data Source. I create the data source and save it, but when i try and activate it...it just keeps running for a long time and gives a timeout error.
    I have created Info Objects and even a Info Cube...i am able to save and activate them without a problem, its only the data sources that give me this problem
    Can somebody throw some light on this issue.
    Thanks
    Ravi.

    Hi,
    can u plz paste ur error here .. It helps us to answer in a better way .
    from ur posted message .. try this ..
    change parameter icm/keep_alive_timeout .. but u need to bounce the system as it not dynamic parameter.
    Thanks

Maybe you are looking for

  • Exporting Final Cut Pro

    I have a 6 minute video that my client wants to proof via an ftp site. They want me to upload the file that they can then download and view on a PC. What export settings and sizings should be used for this to work on the PC and still be a reasonable

  • I need to change my country

    Hi, I've had a trial of creative cloud and am looking to upgrade to a complete account, however when I go to pay it says my country is the UK however I am located in Australia. Please can you help with this?

  • Reg  module pool programming

    Hi, iam developing object on modulepool. can u pls send examples objects on modulepools. thank

  • Connect refused in Exchange Portlet

    Hi, I am trying to config the exchange portlet using URL-services and basic authentication. When I get to create web provider, the following error show up: An error occurred when attempting to call the providers register function. (WWC-43134) An unex

  • Making an Animation

    Hi Eveyone...I installed Adobe Flash CS5.5 a couple of days back...Trying to learn a few basic tricks such as motion tween (messed up things with the bone tool though)..Well the problem is, a friend of mine made a 3D drawing in Solid Works(a modellin