Error when activating Real Time Data Source

Hi..
I created a generic datasource in R/3 system via RSO2, with the Real-time check box enabled in the settings for Generic Delta. I saved the datasource. Replicated it into BW System.
I tried to activate this Real time Data Source in BW, it says "Error when activating Data Source".
I am on BI 7.0, with PI_BASIS: 2005_1_700.
Please let me know if you need any further information.
Thanks,
Sai.

Hi sai,
Can you specify from where you have created your geneic data source like by using DB tabl/view/infoset query.
am

Similar Messages

  • Error when activating the master data

    Hi
    I am getting the following error when activating the master data. when using IDOC as transfer method in transfer rules
    <b>Error while creating table 'EDISEGMENT' entry '/BIC/CIRB0CLASS_NUM_TEXT'</b>
    If anyone came across this kind of problem plz share the solution.
    Regards
    Annie

    Hi
    I solved the problem. AFter changing from PSA to IDOC, i just saved first and activated. it got activated. But thanks for the response.
    There are no short dumps in ST22
    Regards
    Annie

  • Error when creating a new Data Source

    All, This is on BI Publisher 11g.
    When I create a new data source and supply password and then click "Apply" button on top right corner, i get following error
    Error 500--Internal Server Error
    From RFC 2068 Hypertext Transfer Protocol -- HTTP/1.1:
    10.5.1 500 Internal Server Error
    The server encountered an unexpected condition which prevented it from fulfilling the request.
    what am i doing wrong?
    Thank you for reading this post.
    Thanks,
    R

    The schema/user is system and the password is alpha-numeric, upper/lower case, no special char ...
    As a workaround I was able to add the password to the entry in the datasources.xml file and restarted BI Pub.
    The JDBC is working ... I can test the connection successfully ... just wont save the entry
    Edited by: tread on Oct 17, 2011 11:55 AM

  • Error when deploying JDBC file data source for J2EE agent in weblogic

    Hi ODI experts!
    we're using ODI 11g and we've installed J2EE agents on weblogic. We need to configure a connection pool within weblogic for a file data source. Thus, we've added snpsFile.jar (that was shipped with ODI 10g...) to the weblogic config so that we can configure a JDBC connection to file.
    As we validate the config, an error occurs (see below):
    Obviously, driver is found but cannot be deployed.
    ####<4 août 2011 09 h 39 CEST> <Error> <Deployer> <ppdlodi001> <SRVPPDODI_001> <[ACTIVE] ExecuteThread: '2' for queue: 'weblogic.kernel.Default (self-tuning)'> <<WLS Kernel>> <> <c53bfd6260db74bb:-485a441a:13193b42397:-7ffc-0000000000000027> <1312443551325> <BEA-149265> <Failure occurred in the execution of deployment request with ID '1312443550020' for task 'weblogic.deploy.configChangeTask.6'. Error is: 'weblogic.application.ModuleException: '
    weblogic.application.ModuleException:
    at weblogic.jdbc.module.JDBCModule.prepare(JDBCModule.java:290)
    at weblogic.application.internal.flow.ModuleListenerInvoker.prepare(ModuleListenerInvoker.java:199)
    at weblogic.application.internal.flow.DeploymentCallbackFlow$1.next(DeploymentCallbackFlow.java:508)
    at weblogic.application.utils.StateMachineDriver.nextState(StateMachineDriver.java:41)
    at weblogic.application.internal.flow.DeploymentCallbackFlow.prepare(DeploymentCallbackFlow.java:149)
    at weblogic.application.internal.flow.DeploymentCallbackFlow.prepare(DeploymentCallbackFlow.java:47)
    at weblogic.application.internal.BaseDeployment$1.next(BaseDeployment.java:1223)
    at weblogic.application.utils.StateMachineDriver.nextState(StateMachineDriver.java:41)
    at weblogic.application.internal.BaseDeployment.prepare(BaseDeployment.java:367)
    at weblogic.application.internal.SingleModuleDeployment.prepare(SingleModuleDeployment.java:43)
    at weblogic.application.internal.DeploymentStateChecker.prepare(DeploymentStateChecker.java:154)
    at weblogic.deploy.internal.targetserver.AppContainerInvoker.prepare(AppContainerInvoker.java:60)
    at weblogic.deploy.internal.targetserver.operations.ActivateOperation.createAndPrepareContainer(ActivateOperation.java:208)
    at weblogic.deploy.internal.targetserver.operations.ActivateOperation.doPrepare(ActivateOperation.java:98)
    at weblogic.deploy.internal.targetserver.operations.AbstractOperation.prepare(AbstractOperation.java:217)
    at weblogic.deploy.internal.targetserver.DeploymentManager.handleDeploymentPrepare(DeploymentManager.java:749)
    at weblogic.deploy.internal.targetserver.DeploymentManager.prepareDeploymentList(DeploymentManager.java:1216)
    at weblogic.deploy.internal.targetserver.DeploymentManager.handlePrepare(DeploymentManager.java:218)
    at weblogic.deploy.internal.targetserver.DeploymentServiceDispatcher.prepare(DeploymentServiceDispatcher.java:160)
    at weblogic.deploy.service.internal.targetserver.DeploymentReceiverCallbackDeliverer.doPrepareCallback(DeploymentReceiverCallbackDeliverer.java:171)
    at weblogic.deploy.service.internal.targetserver.DeploymentReceiverCallbackDeliverer.prepare(DeploymentReceiverCallbackDeliverer.java:41)
    at weblogic.deploy.service.internal.statemachines.targetserver.AwaitingContextUpdateCompletion.callDeploymentReceivers(AwaitingContextUpdateCompletion.java:164)
    at weblogic.deploy.service.internal.statemachines.targetserver.AwaitingContextUpdateCompletion.handleContextUpdateSuccess(AwaitingContextUpdateCompletion.java:66)
    at weblogic.deploy.service.internal.statemachines.targetserver.AwaitingContextUpdateCompletion.contextUpdated(AwaitingContextUpdateCompletion.java:32)
    at weblogic.deploy.service.internal.targetserver.TargetDeploymentService.notifyContextUpdated(TargetDeploymentService.java:225)
    at weblogic.deploy.service.internal.DeploymentService$1.run(DeploymentService.java:190)
    at weblogic.work.SelfTuningWorkManagerImpl$WorkAdapterImpl.run(SelfTuningWorkManagerImpl.java:528)
    at weblogic.work.ExecuteThread.execute(ExecuteThread.java:201)
    at weblogic.work.ExecuteThread.run(ExecuteThread.java:173)
    Caused By: weblogic.common.ResourceException: com.sunopsis.jdbc.driver.file.a.i
    at weblogic.jdbc.common.internal.ConnectionEnvFactory.createResource(ConnectionEnvFactory.java:263)
    at weblogic.common.resourcepool.ResourcePoolImpl.makeResources(ResourcePoolImpl.java:1193)
    at weblogic.common.resourcepool.ResourcePoolImpl.makeResources(ResourcePoolImpl.java:1117)
    at weblogic.common.resourcepool.ResourcePoolImpl.start(ResourcePoolImpl.java:244)
    at weblogic.jdbc.common.internal.ConnectionPool.doStart(ConnectionPool.java:1110)
    at weblogic.jdbc.common.internal.ConnectionPool.start(ConnectionPool.java:147)
    at weblogic.jdbc.common.internal.ConnectionPoolManager.createAndStartPool(ConnectionPoolManager.java:386)
    at weblogic.jdbc.common.internal.ConnectionPoolManager.createAndStartPool(ConnectionPoolManager.java:326)
    at weblogic.jdbc.module.JDBCModule.prepare(JDBCModule.java:254)
    at weblogic.application.internal.flow.ModuleListenerInvoker.prepare(ModuleListenerInvoker.java:199)
    at weblogic.application.internal.flow.DeploymentCallbackFlow$1.next(DeploymentCallbackFlow.java:508)
    at weblogic.application.utils.StateMachineDriver.nextState(StateMachineDriver.java:41)
    at weblogic.application.internal.flow.DeploymentCallbackFlow.prepare(DeploymentCallbackFlow.java:149)
    at weblogic.application.internal.flow.DeploymentCallbackFlow.prepare(DeploymentCallbackFlow.java:47)
    at weblogic.application.internal.BaseDeployment$1.next(BaseDeployment.java:1223)
    at weblogic.application.utils.StateMachineDriver.nextState(StateMachineDriver.java:41)
    at weblogic.application.internal.BaseDeployment.prepare(BaseDeployment.java:367)
    at weblogic.application.internal.SingleModuleDeployment.prepare(SingleModuleDeployment.java:43)
    at weblogic.application.internal.DeploymentStateChecker.prepare(DeploymentStateChecker.java:154)
    at weblogic.deploy.internal.targetserver.AppContainerInvoker.prepare(AppContainerInvoker.java:60)
    at weblogic.deploy.internal.targetserver.operations.ActivateOperation.createAndPrepareContainer(ActivateOperation.java:208)
    at weblogic.deploy.internal.targetserver.operations.ActivateOperation.doPrepare(ActivateOperation.java:98)
    at weblogic.deploy.internal.targetserver.operations.AbstractOperation.prepare(AbstractOperation.java:217)
    at weblogic.deploy.internal.targetserver.DeploymentManager.handleDeploymentPrepare(DeploymentManager.java:749)
    at weblogic.deploy.internal.targetserver.DeploymentManager.prepareDeploymentList(DeploymentManager.java:1216)
    at weblogic.deploy.internal.targetserver.DeploymentManager.handlePrepare(DeploymentManager.java:218)
    at weblogic.deploy.internal.targetserver.DeploymentServiceDispatcher.prepare(DeploymentServiceDispatcher.java:160)
    at weblogic.deploy.service.internal.targetserver.DeploymentReceiverCallbackDeliverer.doPrepareCallback(DeploymentReceiverCallbackDeliverer.java:171)
    at weblogic.deploy.service.internal.targetserver.DeploymentReceiverCallbackDeliverer.prepare(DeploymentReceiverCallbackDeliverer.java:41)
    at weblogic.deploy.service.internal.statemachines.targetserver.AwaitingContextUpdateCompletion.callDeploymentReceivers(AwaitingContextUpdateCompletion.java:164)
    at weblogic.deploy.service.internal.statemachines.targetserver.AwaitingContextUpdateCompletion.handleContextUpdateSuccess(AwaitingContextUpdateCompletion.java:66)
    at weblogic.deploy.service.internal.statemachines.targetserver.AwaitingContextUpdateCompletion.contextUpdated(AwaitingContextUpdateCompletion.java:32)
    at weblogic.deploy.service.internal.targetserver.TargetDeploymentService.notifyContextUpdated(TargetDeploymentService.java:225)
    at weblogic.deploy.service.internal.DeploymentService$1.run(DeploymentService.java:190)
    at weblogic.work.SelfTuningWorkManagerImpl$WorkAdapterImpl.run(SelfTuningWorkManagerImpl.java:528)
    at weblogic.work.ExecuteThread.execute(ExecuteThread.java:201)
    at weblogic.work.ExecuteThread.run(ExecuteThread.java:173)
    Thanks in advance for your help.
    Fredd.

    Hi!
    Some information: I found this website [http://gerardnico.com/doc/odi/webhelp/en/refmanual/connexion/jdbcdriversample.htm] explaining that the 'old' snpsFile.jar driver was deprecated and that a new one was existing... but I can't find it anywhere!
    My questions are:
    - Can the problem (original problem. See initial post) be due to driver version?
    - Does someone know where to find this 'new' driver?
    Thanks for your help.
    Fredd.
    Edited by: fredd2 on Aug 5, 2011 10:04 AM
    Edited by: fredd2 on Aug 5, 2011 10:07 AM

  • Short Dump Error when we check the Data Source in RSA3

    Hi. Experts.
                       We are using  Data Source 2Lis_13_VDITM and when we try to check in RSA3
    for recorts it is going short dump. I am just providing the error information.
                  An exception occurred that is explained in detail below.
    The exception, which is assigned to class 'CX_SY_IMPORT_MISMATCH_ERROR', was
    not caught in procedure "MCEX_BW_LO_API" "(FUNCTION)", nor was it propagated by a RAISING
    clause.
                   Since the caller of the procedure could not have anticipated that the exception would occur, the current program is terminated.
    The reason for the exception is:
    When importing the object "MC13VD0ITM_TAB", the component no. 96 in the dataset has a different type from the corresponding component of the target object in the program "SAPLMCEX".
                        The data type is "N" in the dataset, but "X" in the program.
            Since I don't have much knowledge in ABAP, Please help me out regarding this issue. Suggest me possible solution and let me know if you need any further information.

    Hi,
    Try adding a TRY CATCH block in the extractor function module to catch the exception.
    For the type mismatch you will have to make sure the variables are of same type or they can be type converted.
    Use the debug mode of RSA3 to find what data is being exchanged.
    You will need to change your extractor to avoid the dump.. or ensure that the value being passed to that object is correctly being passed(correct format)
    Cheers
    Kiran

  • "logon failed" error when connecting to XML data source

    I have an HTTP source that generates XML. I have a schema that describes this XML. If I save the XML to a local file it works fine as a data source with the schema. However, when I try to access the same data via HTTP, I get the following error:
    Logon failed.
    Details: Cannot open file
    Server returned HTTP response code: 401 for URL: http://localhost:8004/report.xqy?Validate%20XML=0&Use%20WS-Security%20Config%20File=&WS-Security%20File%20Location
    However, if I enter that URL in a browser it works fine. And even if I disable security on the HTTP source, it produces the same error.
    I have tried a local HTTP source as well as the same source running on Amazon EC2.
    Any ideas?

    Thanks for the help.
    I am new to crystal reports, so I do not understand your suggestion. Both of those points look like they are related to Adobe Flash.
    I am trying to create a report using the standard report creation wizard:
    1) new, standard report
    2) create  new connection
    3) xml and web services
    4) xml data source
    When using the sample from the SAP site, it works fine:
    http://resources.businessobjects.com/support/downloads/samples/cr/customer_db/customer.xml
    When connecting to my source from a browser, it works fine. When using my source as a local document instead of over http it works fine. It is only when connecting to my source over http that I get the failed logon error.
    Kelly

  • Error when activating the copied data elem.:Medium key word in language DE

    Hello,
    I have copied data element dats and created a z data element for myself and added the text description.
    Am getting following error.Please guide me for the same-
      Check data element ZDATS1 (C04EAPHALES/15.01.10/16:37)
      Medium key word in language DE: length 18 > maximum length 10
      Data element ZDATS1 is inconsistent
    Please help me resolve the issue.
    Thanks.
    Edited by: sanjivrd on Jan 15, 2010 12:21 PM

    Hi,
    It won't take from domain it always takes from data element... data element only stores the semantic information... Try to change the domain description to someother and then activate both domain and dataelement...and also table and then see...It will take from data element only.
    Regards,
    Nagaraj

  • Real-time data acquisition for HR datasources

    Dear Experts,
    I have a couple of questions on real-time data acquisition...
    a) Can you tell me if any standard HR datasources support real-time data acquisition?
    b) Can we apply real-time data acquisition for generic datasources? If yes, is there any difference in the process compared to working with business content datasources when using real-time data acquisition?
    Hope you can provide some answers...as always, points will be awarded for answers that effectively address the questions.
    Thanks a bunch.
    K

    Hi Karthik,
    a)The decision to go for SAP remote cube depends on the volume of data and the frequency of queries. It is advicible not to use the SAP remote cube if the data volume is high and also, the query frequency is high. Which module in HR are yuo implementing? In HR, the data volume is generally high. So, if you go for loading the data from R/3 every 3 hours, you are asfe as long as the loading volumes are less. For example, for implementing Time management, I would not advice frequent loads as it is time consuming process. So, make a decision based on the above mentioned lines. If the data volume is not high, I would prefer the SAP ermote cube as it will reduce managing the loads.
    b)I mentioned FM extractor just for the sake of control of data selection. You can even go for view/table extractor.
    Hope this helps.
    Thanks and Regards
    Subray Hegde

  • Error - When Activating Generic Data source in BW DEV

    Hi,
    I am working in BI 7.0
    I created a Generic Data source
    I had checked in RSA3 , it is showing records too.
    i replicated the Datasource in BW
    Datasource is available in BW
    Then I try to activated the Datasource
    because
    when i try to create tranformation for the data source then
    It will show this message (and transformation is not created)
    <b>Source RSDS ZCOPC_C10 GDVCLNT110 does not exist</b>
    My Data source name is ZCOPC_C10 for source system GDVCLNT110
    then I will try to Activated the Data source First
    then also it will show the Error like this
    Post Processing/Checking the Activation for DataSource ZCOPC_C10 GDVCLNT110
    <b>&
    Error when activating DataSource ZCOPC_C10                     GDVCLNT110</b>
    and the
    DataSource ZCOPC_C10 (GDVCLNT110) could not be activated
    can some body help me to solve this problem
    sure you can get the points
    Thanks
    prasanna

    Hi Oscar,
    The Data source is not Active in BW, I had checked in Data source defination.
    when I activate the Data source .
    <b>I received the Message / Error</b>
    Activation of Objects with Type DataSource
    Saving Objects with Type DataSource
    Internal Activation (DataSource )
    Preprocessing / Creation of DDIC Objects for DataSource ZCOPC_C11 GDVCLNT110
    Save DDIC object /BIC/CAZCOPC_C000020000001 (Application Structure for ZCOPC_C11 GDVCLNT110)
    Save DDIC object /BIC/WCAZCOPC_C000020000001 (Application Structure for ZCOPC_C11 GDVCLNT110 - T)
    Save DDIC object /BIC/CAZCOPC_C00002000 (Segment Structure for ZCOPC_C11 GDVCLNT110 - Appli)
    Save DDIC object /BIC/CCZCOPC_C000020000001 (Character Structure for ZCOPC_C11 GDVCLNT110)
    Save DDIC object /BIC/WCCZCOPC_C000020000001 (Character Structure for ZCOPC_C11 GDVCLNT110 - Tab)
    Save DDIC object /BIC/CCZCOPC_C00002000 (Segment Structure for ZCOPC_C11 GDVCLNT110 - Chara)
    Save DDIC object /BIC/B0001051 (PSA for ZCOPC_C11 GDVCLNT110 Itemization)
    Writing of object catalog entries (TADIR)
    Object TABL /BIC/CAZCOPC_C000020000001 written to the ObjectCatalog (TADIR)
    Object TTYP /BIC/WCAZCOPC_C000020000001 written to the ObjectCatalog (TADIR)
    Object TABL /BIC/CAZCOPC_C00002000 written to the ObjectCatalog (TADIR)
    Object TABL /BIC/CCZCOPC_C000020000001 written to the ObjectCatalog (TADIR)
    Object TTYP /BIC/WCCZCOPC_C000020000001 written to the ObjectCatalog (TADIR)
    Object TABL /BIC/CCZCOPC_C00002000 written to the ObjectCatalog (TADIR)
    Activate all Dictionary objects ( 6 ):
    All DDIC objects have been activated / deleted
    <b>Post Processing/Checking the Activation for DataSource ZCOPC_C11 GDVCLNT110
    Error when activating DataSource ZCOPC_C11                     GDVCLNT110</b>Deletion of Unused Dictionary Objects ( 6)
    All DDIC objects have been activated / deleted
    Deletion of Object Catalog Entries (TADIR)
    Object TABL /BIC/CAZCOPC_C000020000001 deleted from TADIR
    Object TTYP /BIC/WCAZCOPC_C000020000001 deleted from TADIR
    Object TABL /BIC/CAZCOPC_C00002000 deleted from TADIR
    Object TABL /BIC/CCZCOPC_C000020000001 deleted from TADIR
    Object TTYP /BIC/WCCZCOPC_C000020000001 deleted from TADIR
    Object TABL /BIC/CCZCOPC_C00002000 deleted from TADIR
    All unused DDIC objects have been deleted
    Thanks
    prasanna

  • Error "cannot load request real time data targets" for new cube in BI 7.

    Hi All,
    WE have recently upgarded our SCM system from 4.1 to SCM 7.0 which incorporated BI 7.0.
    I am using BI 7.0 for first time and ahve the following issue:
    I ceated a new infocube and data source of flat file and succesfully created transformation, and Data Transfer Process. Everything looked fine. I added flat file and checked preview and could see data. Now when I start job to load data in infocube the follwing error is shown "cannot load request real time data targets". 
    I checked cube type in setting in infcune is shows as Standard.  When I doube clicked on error the following message showed up
    You are trying to load data into a real-time InfoCube using a DTP.
    This is only possible if the correct load settings have been defined for the InfoCube.
    Procedure
    In the object tree of the Data Warehousing Workbench, call Load Behavior of Real-Time InfoCube from the context menu of the InfoCube. Switch load behavior to Transactional InfoCube can be loaded; planning not allowed.
    I did not understand what it is meant and how to set changes. Can someone advice and follow me through.
    Thanks
    KV

    Hi Kverma,
    Real-time InfoCubes can be filled with data using two different methods: using the transaction for entering planning data, and using BI staging, whereby planning data cannot be loaded simultaneously. With Real time cube you can select the method you want to use for update as
    Real Time data Target can be loaded With Data; Planning not allowed &
    Real Time data Target can be Planned; Data loading not allowed
    You can change this behaviour by right clicking on cube and selecting Change real time load behaviour and select first option. You will be able to load the data then
    Regards,
    Kams

  • Error when activating Datasource in Source System

    Dear All,
    I am gettting this error when activating data source in source system(RSA5).
    1)DataSource 0NETWORK_CUST_ATTR; switch to package PS_IS_EHP3_SFWS_SC is off.
    2) The extract structure DTTV_REC_53 of the DataSource 0TV_REC_531 is invalid
    Message no. R8444
    Diagnosis
    An invalid extract structure has been assigned to the DataSource. An extract structure has to be active in the DDIC. It cannot be a view, since the customer is not able to add append fields for filling in the customer exits to a view.

    Hi
    To use this datasource, you need to install EHP3 package.
    http://help.sap.com/saphelp_nw70/helpdata/EN/b0/7b6c759c494ffcbf821d2442687888/frameset.htm
    Chandu.

  • Error when activating datasource

    Hello,
    I am trying to create a new transaction data datasource assigned to a flat file source system.  When i hit the check i get no errors but when i try to activate the activation fails with errors.  The error Message No. RSO404 says "Error when activating Datasource".  We are on BW 7.0 stack 12.
    Has anyone seen this?  It is hard to troubleshoot it when i get no real errors to go on.  It is a very simple three field datasource with one characteristic, one key fig, and an additional currency characteristic.  The key fig is in external format and references 0CURRENCY.  Does anybody have any insight?
    Thanks,
    TMS

    Hi TMS,
    I am facing exactly the same problem here...
    transactional datasource for flatfile gives me errors DO313, RSO404 and then RSDS106
    We are on:
    SAP_ABA            700               0011  SAPKA70011
    SAP_BASIS     700               0011  SAPKB70011
    PI_BASIS            2005_1_700  0011  SAPKIPYJ7B
    SAP_BW            700               0012  SAPKW70012
    FINBASIS            600               0008  SAPK-60008INFINBASIS
    BI_CONT            703               0005  SAPKIBIIP5
    SEM-BW            600               0008  SAPKGS6008
    Did you manage to find a solution?

  • Problem with Real Time Data Acquisition

    Hello,
    I tried to built a RDA data flow according to the SAP-Help and this blog (/people/kamaljeet.kharbanda/blog/2006/11/13/real-time-data-acquisition-bi2004s), but I ran into a problem. Everything seems to be fine, but even though the daemon is running no new data is added to the DSO. I'll explain in detail what steps I took and hope someone can pinpoint my error(s).
    System is BW4 on SPS 17. My datasource as well as my DSObject are in the same system
    1.) Created a transparent table with a timestamp-field for generic delta as a local object.
    2.) Created a generic datasource from view/table (used the above table), marked it as realtime-enabled and used timestamp as delta-specific field.
    3.) Replicated the metadata for my datasource and activated it.
    4.) Created a Data Storage Object with infoobjects corresponding to the table fields
    5.) Created a transformation between DataSource and DSO
    6.) Created a DTP from Data Source to DSO. The type is realtime-DTP.
    7.) Created an InfoPackage for Delta-Initial Load. Update mode: Initialize Delta Process -> Initialization with Data Transfer.
    8.) Then I changed my DTP (step 6) to "normal DTP" and executed it. Then activated the data in the DSO. The datasets from the table (which i filled with an ABAP-program) were successfully loaded and activated. So now my DSO contains 5 datasets.
    9.) Changed the DTP back to "realtime DTP"
    10.) Created an InfoPackage for RDA. Adapter is set to "Realtime extraction from SAP System"
    11.) Created a daemon, assigned the IP for RDA and the DTP to the daemon and started it up.
    12.) Executed my ABAP program to added 3 datasets to the table which is used as a datasource.
    So now in my source table there are 8 datasets, but the daemon won't load them. In my RDA-Monitor (trx rsrda) it shows my daemon, the assigned IP and the assigned DTP (status green). Under the IP there's one yellow request (that should be right according to the help), but unlike the picture in the blog (link see above) there's no request under the DTP.
    Job overview (sm37) shows an active batch job and some jobs that are ready (every 10minutes a new batch starts and the old one is closed). the job log shows, that the daemon runs, but loads no data...
    i think it might be a problem with the delta extraction somewhere, because the intial load works fine and there are no errors anywhere.
    Edit: perhaps my settings in the data source unter tab: extraction are wrong?
    Delta process: AIE After Images By Extractor (can't change that)
    Direct Access: Allowd
    real Time: Real-Time DA Supported
    Adapter: realtime data extraction (also tried Access to SAP Data through Service API which seems senseless)
    Data Format: Fixed Lenght
    Anyone can explain what those fileds mean and which one i should use?
    Edited by: Christoph Jender on Apr 10, 2008 2:48 PM
    Edited by: Christoph Jender on Apr 10, 2008 3:33 PM

    Hi again, another thing i'd like to try is to just active a business content datasource which is realtime-enabled and has the ability to dynamically add some values so that i can test if it works. but i have no idea what business content would be applicable here, any ideas on that perhaps?

  • "Error when activating update rule" after changing ODS key

    Hi Gurus,
    I have problem when transporting in productive system an existing ODS & the update rule linked. The message is  :  "Error when activating update rule".
    This is when i move one field from the key to a simple field of the ODS.
    Before transporting requests, i deleted data in the ODS by right-click on ODS --> Request tab --> select the request --> click on delete....
    I don't understand because i did the same in Qual environment and everything is OK : transport request & ODS loading...
    Please help.....

    Hello,
    Please try to check/perform the following steps:
    1- You must ensure for the update rule have mapping between infosources
       and ODS/CUBEs correct, this means that also these objects need
       to be of the same version in the source system of the transport as in
       target system.
    2- Please ensure all infoobjects in the cube are active.Please
       reactivate all the objects again in source system.
    3- Ensure all infoobjects are found in the communication structure.
    4- Please try to collect the update rule via BW transport connection
       with the grouping option "only necessary objects" and then transport
       those objects again.
    Could you please reactivate the infocube and its update rules again
    and then create a new transport for this infocube and transport it
    again.
    Best regards,
    Paula Csete

  • 0FIA_DS11:Error when activating DataStore Object

    Hi,
    We have activated the datasources related FI_AA and replicated in to a new BI 7 system.
    Currently we are able to activate the asset accounting related targets like 0FIAA_C11, 0FIAA_c12 and 0FIAA_C13
    however not able to activate DSO's 0FI_DS11, 0FI_DS12 and 0FI_DS13.
    We have replicated the datasources 0FI_AA_11 and 0FI_AA_12  in 3.x but while installing the DSO 0FI_DS11 and 0FI_DS12, facing below given  error.
    " DataSource 80FIA_DS11 does not exist in source system BIDCLNT600 of version A     
    Error when creating the export DataSource and dependent Objects
    Error when activating DataStore Object 0FIA_DS11 "
    We have checked the mysource system settings and activated the Data Marts respectively but still we are not able to activate.
    System is BI 7 with EHP4 ehanced packege.
    Please suggest.
    Thanks & Regards,
    SMS

    Hi Jonas,
    I understand the procedures and had followed all the procedures.
    We had resolved the issue and issue was with the IP in sm59.
    It is cleared.
    Thanks for the reply.
    Regards,
    S M S

Maybe you are looking for

  • PSE10: Table keystroke sequence to clear a layer?

    Hi, I'm trying to use CTRL+A;Delete;CTRL+D key combination in order to clear the content of the active layer. When I'm doing it manually via the keyboard, the results are as expected. But when I feed this and a sequence of keystrokes to one of my Wac

  • Under what conditions does a STB reset itself?

    Just curious. My 6416 seems to do it when I am in VOB or when I'm trying to get it to process too many things at once. Any of you techies know?

  • Prime Infrastructure with Mesh Networks

    My customer has a Mesh network at a sea port.  They have about 15 mesh access points at each location with two controllers each.  Would Prime Infrastructure work with a mesh deployment?  There is a lot of interference and sometimes the performance of

  • Probably an easy fix....

    hey there all, I have a movie clip that has a few layers and a button on one of the layers. on the button I have this code: on(rollOver){ gotoAndPlay("over"); on(rollOut){ gotoAndPlay("out"); Now, this mc is on the Main timeline and has an istance na

  • How much partition space should I have for the windows side?

    I need to set up a partition to run windows on my mac.  what is the minimum space I should use for the partition so that windows doesn't get bogged down?