Data source on top of virtual cube s_r_infoprov  expection

hi experts,
i have generated a export datasource on virtual cube. The virtual cube is based on a function module.
When I display data in virtual cube it shows all the data .When I extract the data from it through DTP , it works well.
But when i am using data source(exported data source) to pull the data from the virtual cube it is not pulling the data.
It is throwing an exception at the following place.
CALL FUNCTION 'RSDRI_INFOPROV_READ'
       EXPORTING
                                                        "begin qyh030513
         i_infoprov             = l_infoprov  "'/CPMB/ASFIN'
*        i_infoprov             = do_util->do_appl_info_m-MULTIPROV  "'/CPMB/ASFIN'
                                                          "end qyh030513
         i_th_sfc               = lth_sfc
         i_th_sfk               = lth_sfk
         i_t_range              = lt_range
         i_rollup_only          = space
         "i_use_db_aggregation   = abap_true "abap_true "RS_C_TRUE
         i_use_db_aggregation   = IF_DB_AGGREGATE "abap_true "RS_C_TRUE
         i_use_aggregates       = abap_true "abap_true "RS_C_TRUE
         i_packagesize          = i_packagesize
         i_authority_check      = space
         i_currency_conversion  = space
       IMPORTING
         e_t_data               = et_data
         e_split_occurred       = e_split_occurred
         e_end_of_data          = e_end_of_data
         e_stepuid              = l_stepuid
       CHANGING
         c_first_call           = c_first_call
       EXCEPTIONS
         illegal_input          = 1
         illegal_input_sfc      = 2
         illegal_input_sfk      = 3
         illegal_input_range    = 4
         illegal_input_tablesel = 5
         no_authorization       = 6
         illegal_download       = 8
         illegal_tablename      = 9
         OTHERS                 = 11.
When we go inside this function module in debugging the below code will be there
STATICS: s_r_infoprov   TYPE REF TO cl_rsdri_infoprov.
   CLEAR: e_t_msg.
   IF i_clear = rs_c_true.
     CLEAR s_r_infoprov.
     RETURN.
   ENDIF.
IF s_r_infoprov IS NOT INITIAL AND c_first_call = rs_c_true.
*   nested call of RSDRI_INFOPROV_READ
     MESSAGE e882(dbman) RAISING illegal_input.
   ENDIF.
Problem is with the s_r_infoprov value.  When I execut with data source in RSA3 then  the value is {O:222*\CLASS=CL_RSDRI_INFOPROV}
and when I executed directly the virtual cube the value for s_r_infoprov is initial.
So because of this value {O:222*\CLASS=CL_RSDRI_INFOPROV} the following code got executed and throwing an exception.
Experts please guide how to solve this. will data source on top of virtual cube wont work?
thank you

Hi Sander,
Thank you for the reply.
I am trying to understand why an virtual cube is not working if it is generated as an exported  data source and it is working perfectly if directly connected as a source to DSO or Cube?
Problem is with the s_r_infoprov value. It is getting populated differently , that is why there is an exception which using data source(virtual cube generated as exported data source).
If this functionality is not going to work then why SAP has given an option to generate an exported DS of virtual cube.
thank you for all giving some valuable inputs on this. Some how some thing is missing it seems....

Similar Messages

  • Source system Assignment for Virtual Cube

    HI,
    I transported infosource and transfer rules to production, but source system is not transported. if i am right i think i need to assign source system(prd) manually. but in prd all objects are not changeble, then how can i assign the source system to infosouce.
    Regards
    Rajini

    HI,
    I transported infosource sucessfully to production , but when i am trying to transport transfer rules , giving warning source system is not transportable. this is happening with remote cube only. all other normal transports are fine.
    i think there is a special way to transport TR for remote cube. pls. help me.
    Rajini

  • Data source to Virtual Cube

    Hello friends,
    I am having a Virtual Cube with services based on Functional Module.
    How can i find out from where data is comming into this virtual cube.
    I would like to see the source data which is comming into this Virtual Cube?
    Thanks
    Tony

    Hi Tony,
    For Virtual Infoprovider with services, there will be a FM in BW system itself. All the connection and selection of data is done in the FM. So, You do not have a separate DS for them. Just look into the FM for the logic of connection and extraction.
    Thanks and Regards
    Subray Hegde

  • Not able to fetch the data by Virtual Cube

    Hi Experts,
    My requirement is I need to fetch the data from Source System (From Data base table) by using Virtual Cube.
    What I have done is I have created Virtual Cube and created corresponding Function Module in Source System.
    That Function MOdule is working fine, if Data base table is small in Source System.But If data base table contains huge amount of data (millions of record), I am not able to fetch the data.
    Below is the code I have incorporated in my function module.
    c_th_mapping TYPE CL_RSDRV_EXTERNAL_IPROV_SRV=>TN_TH_IOBJ_FLD_MAPPING.
      DATA:
        l_s_map TYPE CL_RSDRV_EXTERNAL_IPROV_SRV=>TN_S_IOBJ_FLD_MAPPING.
      l_s_map-iobjnm = '0PARTNER'.
      l_s_map-fldnm  = 'PARTNER'.
      insert l_s_map into table l_th_mapping.
    create object l_r_srv
        exporting
           i_tablnm              = '/SAPSLL/V_BLBP'
          i_th_iobj_fld_mapping = l_th_mapping.
      l_r_srv->open_cursor(
        i_t_characteristics = characteristics[]
        i_t_keyfigures      = keyfigures[]
        i_t_selection       = selection[] ).
       l_r_srv->fetch_pack_data(
        importing
          e_t_data = data[] ).
      return-type = 'S'.
    In the above function Module,Internal table L_TH_MAPPING contains Info Objects from Virtual Cube and corresponding field from Underlying data base table.
    The problem where I am facing is, in the method FETCH_PACK_DATA, initially program is trying to fetch all the recordsfrom data base table to internal table.If Data base table so lagre, this logic is not working.
    So would you please help me how to handle these kind of issues.

    Hi Experts,
    My requirement is I need to fetch the data from Source System (From Data base table) by using Virtual Cube.
    What I have done is I have created Virtual Cube and created corresponding Function Module in Source System.
    That Function MOdule is working fine, if Data base table is small in Source System.But If data base table contains huge amount of data (millions of record), I am not able to fetch the data.
    Below is the code I have incorporated in my function module.
    c_th_mapping TYPE CL_RSDRV_EXTERNAL_IPROV_SRV=>TN_TH_IOBJ_FLD_MAPPING.
      DATA:
        l_s_map TYPE CL_RSDRV_EXTERNAL_IPROV_SRV=>TN_S_IOBJ_FLD_MAPPING.
      l_s_map-iobjnm = '0PARTNER'.
      l_s_map-fldnm  = 'PARTNER'.
      insert l_s_map into table l_th_mapping.
    create object l_r_srv
        exporting
           i_tablnm              = '/SAPSLL/V_BLBP'
          i_th_iobj_fld_mapping = l_th_mapping.
      l_r_srv->open_cursor(
        i_t_characteristics = characteristics[]
        i_t_keyfigures      = keyfigures[]
        i_t_selection       = selection[] ).
       l_r_srv->fetch_pack_data(
        importing
          e_t_data = data[] ).
      return-type = 'S'.
    In the above function Module,Internal table L_TH_MAPPING contains Info Objects from Virtual Cube and corresponding field from Underlying data base table.
    The problem where I am facing is, in the method FETCH_PACK_DATA, initially program is trying to fetch all the recordsfrom data base table to internal table.If Data base table so lagre, this logic is not working.
    So would you please help me how to handle these kind of issues.

  • Can anyone explain how the LO data source flows from ECC to BI up to cubes?

    Dear all,
    Can anyone explain step by step (HOW THE DATA FLOWS)....By taking 2LIS_11_VAHDR  from SD , starting from Activate Data source in ECC  to Info Cube in BI7.0. Please this will be helpful -
    Thanks for the advance answers...
    Edited by: harishk.225 on Dec 23, 2011 9:56 AM

    Hi Harish,
    First go to RSA5 in ECC  select your datasource 2LIS_11_VAHDR and activate it.
    After activatin check in RSA6 whether the datasoure activated properly or not.
    Then log on  to BI sytem Select datasource and click on replicate.
    Then Create Infocube infopackage ,Transformation, DTP ...entire flow. But dont schedule it because there is no data in SETUP Tables.For full load first we need to run statistical set up , so that data should come to SETUP Table.
    Now go to ECC  enter T.CODE OLI7BW to filled up the SETUP Table  for datasource 2LIS_11_VAHDR.
    It will ask  you for Run Name etc give the run name and time limit and execute it.
    If you get some error then first delete the SETUP Table by using Transaction Code LBWG.
    In LBWG it will ask you for applicatiion number , give 11 i.e for sales datasourc and execute it.
    Data will get deleted from SETUP Table. To check data is deleted or not  in SET UP Tabe go to database table MC11VA0HDRSETUP. Always remember the name of SETUP Table will be Extract Structure followed by SETUP.
    EX : If extract structure name is MC11VA0HDR then name of the SETUP Table will MC11VA0HDRSETUP.
    If data gets deleted then again go to 0LI7BW and run the Statistical Setup then agan check in MC1VAOHDRSETUP table contains data or not.
    Then trigger the Infopackage and DTP in BI.  Above steps was for full load.
    Now to Load deltas follow below steps.
    First Create INIT at BI side which enable delta at ECC Side.
    Then go to ECC and execute T.CODE LBWE.
    LBWE is LO Cockpit Work Bench. There we have folowng functions 1.Maintan Datasource. 2.Maintain Extract Structure. 3. Job  Control.  4.Delta Type 5.. Activate/deactivate.
    If you want to add any new field to your datasource then to to Maintain Extract Strucutere remeber before adding or modififying datasource first we should deactivate datasource by using 5th function  Activate/Deactivate.
    Then select DELTA TYPE in delta type we have four types of delta. 1. Delta Queue (RSA7) 2. Queued Delta (LBWQ)
    3. Unserialized (SM13) , 4. Seriaized (SM13)
    If you select Delta queue then deltas drectly come to RSAT.
    if you select Queued Delta  deltas goes to LBWQ then you have perform V3 job by using Job Control  function to move data from LBWQ to RSA7. Remeber to BI  data  always goes from RSA7 only.
    If you select Unserealized then delta goes to SM13 again you have to perorm V3 Job to move deltas in RSA7.
    Serialize update is not in used in Lo-cockpit.
    There is a differerence between  DELTA QUEUE , QUEUED DELTA, UNSERIALIZE UPDATE. Pls find the differences on net.
    Then select your delta type and create deltas dtp in BI and start loading .
    Hope this help you.
    In RSA7 we have two tables delta and repeat delta, to know the functionality of this please search for the documents on  net.
    Regards,
    Asim

  • Creation of Generic Data source using function module based on the program which was used to created ABAP report

    Hi,
    We have a requirement to create a BI report based on plant maintenance report. The plant maintenance report is based on a ABAP program with complex logic. My question is i want create a Generic Datasource using Function module and can I include the logic(Abap Program) that is used for plant maintenance report in the function module? Please share your thoughts.
    Thanks,
    Ravi

    Hi,
    Step1-Create a table structure which you need to create same as the fields you require in you data source.
    Step 2-create a custom abap program and inside that call the Client abap program with the selections as required and save the result data in some table
    Step 3-create infoset query.In the infoset query give your table structure name and  program name which you developed.
    Step4-create data source on top of that query

  • SAP HANA as a data source for Analysis Services

    I tried to use SAP's .net provider for HANA as a data source for an Analysis Services cube but to no avail.   I can connect, create a named query and preview data from the data source view designer but when SSAS actually tries to run the query,
    it wraps the named query in another query and produces a syntax error.  
    For example,  if the named query is "select * from ENTHANA.MARA", 
    SSAS sends a query like this "SELECT [test].* FROM (select * from ENTHANA.MARA) AS [test]"   (Test is the name of the named query).
    I know SSAS wraps queries and that's fine - except that brackets are not a valid way to quote an identifier in HANA.  HANA uses double quotes like Oracle. 
    Is there any setting in SSAS that can affect this behavior?   Will HANA ever be a fully supported data source for SSAS?  If so, when?

    Hi David,
    According to your description, you are creating a SQL Server Analysis Services project, now what you want is using SAP HANA as the data source, right?
    SSAS support many type of data source. However, as you can see on the link below, SAP HANA data source was not listed on that link. So this type of data source is not supported in current version of SSAS. Microsoft will update that document when it is supported.
    http://msdn.microsoft.com/en-IN/library/ms175608.aspx
    Thank you for your understanding.
    Regards,
    Charlie Liao
    TechNet Community Support

  • Infospoke with two data source

    Hi all,
    is there a way in an Infospoke to include infoobjects belonging to two different data source?
    tks,
    Federico

    Hello Frederico,
    as far as I know you can only use one data source of type InfoObject, basic cubes or ods in an infospoke. Perhaps Analyses Process Designer (APD) may help you to achieve your goal. You can use it to join your InfoObject into a new one, or create a transactional ODS with the desired data. These new objects can than be used in an infospoke. And you can use a process chain to put all required steps in the right order.
    Kind regards,
    Jürgen

  • Data source Infocubes for standard reports

    Hello Everybody,
    Can you please assist - how to understand which data source will feed - which info-cube - which standard reports?
    I am planning to activate standard reports in various module. How do I decide corresponding data source and info-cube relevant to these reports?
    Thanks for your help.
    TS

    Hi ,
    If you know the standard query name , then you can get the data provider name in the " information " tab ( bex Analyser). Using the infoprovider go to Business content > give , in data flow before and then install the infoprovider and all the relavant objects.Then go to RSA1 and check the infoprovider, right click on the infoprovider >display data flow
    Hope this helps.
    Thanks and regards

  • Processing data base failed: Cannot connect to data source

    Hi,
    I have deployed my SSAS project to an Analysis services database on a SQL 2014 server instance using the "Development" build configration in Visual Studio. In order to avoid users accessing the development instance of the cube, I've created
    an additional "Release" build configuration and deployed my project to a different database on the same server. I've also created a production copy of the data source and changed the data source configuration of the production cube to point to this
    one.
    I've provided the same domain service account on the "Impersonation information" tab of the data source for both the development as well as the production instance of the cube. This account has also been granted identical permissions for both data
    sources. While everything works fine with the development database, processing the production database fails with an error message saying that it cannot connect to its respective data source.
    I'm using VS Ultimate 2013.4, SSDT 12.0.50318.0. Any hints pointing me in the right direction are appreciated.
    Many thanks
    Andreas

    Hi Andreas,
    According to your description, your production cube can't connect to data source when processing. Right?
    In this scenario, if both cube data sources are pointing to same source databases, both connection should work proerly. Since you created a copy of data source for your production cube, please make sure your type the right server name and test connection
    in your project. If you can connect successfully, I suggest you only keep this data source for your production project. If the issue still persists, I suggest you import the development database into a new project and save this one as your production database.
    If you still have any question, plesse provide some more information and share some screenshots.
    Best Regards, 
    Simon Hou
    TechNet Community Support

  • Issue with data source after deploying

    We are experiencing an issue with our data source after deployment of a cube. On the datasource properties in Visual Studio 2012, we have the max connections set to 0 before the deployment. Once the cube is deployed, I can navigate to the <name>.0.ds.xml
    file and open it and see that the <MaxActiveConnections>0</MaxActiveConnections> is indeed set to 0. At some point over the next couple days, a process of the cube or some other action causes that value to get updated to some number too large to
    be converted to an int, and makes the datasource invalid. At that point we cannot view the datasource properties in SSMS, we cannot open the cube project from Visual Studio, and we’ve even had failures when trying to process the cube.  Is there a config
    somewhere that would cause this value to get overwritten, or some other behind the scenes process that we can look at?
    Our server information is:
    Microsoft SQL Server 2012 (SP1) - 11.0.3153.0 (X64)
                    Jul 22 2014 15:26:36
                    Copyright (c) Microsoft Corporation
                    Enterprise Edition: Core-based Licensing (64-bit) on Windows NT 6.2 <X64> (Build 9200: ) (Hypervisor)
    Chad Dotzenrod SWC | TECHNOLOGY PARTNERS 1420 Kensington Road, Suite 110 Oak Brook, Illinois 60523-2144 http://www.swc.com

    Typically you would import the metadata from the source location and either use that location as the data source (and so not need to redeploy), or deploy it to a separate target location.
    The replace action is destructive as you've found, and effectively performs a drop table followed by create table. Hence any data in the table is lost.
    If you just want the Control Center Manager to correctly display that the table is deployed, try setting the action to "Upgrade". This will try to upgrade the deployed object to match the definition in OWB, but as the two are identical this will result in no changes. However, it will update the deployment records to indicate that the object is deployed.
    Nigel.

  • Generic Data source on Function module

    Hi All,
    We are working on generic data source based on a function module for the first time and we do not know how to start with.
    Our requirement is to fetch mode of communication field from SRM. We have identified a function module in SRM which does this job. But when we try to create a generic data source on this function module it is throwing an error.
    Are the import parameters for a general function module and function module for a BI generic data source different?
    If so could someone please guide us in creating the function module so that we can create a generic data source on top of it.
    The function module should serve the same purpose as the function module BAPI_BUPA_ADDRESS_GETDETAIL.
    Thanks,
    Shravani

    Shivani,
    My suggestion would be creating Z-program to store the values in a custom table and use this for extraction.
    Further if still want to use the FM itself, post the error details you are getting.. or try to analyse the FM and find the base table so that you can directly use in extraction.
    Hope this helps,
    anil

  • Joint Venture Accounting Data Sources

    Hi All,
    I am going to work on Joint venture accounting. I need the data sources which we use for cubes 0JVA_C01, 0JVA_C02.
    I have checked in SAP HELP.But, i didn't get  complete info about the cubes.
    Please suggest me......If you have any documents, plzzz send me.My id is [email protected]
    Thanks,
    Kameswari.

    Hi
    Thanks for your reply.
    I have assigned points.
    Here i ahve a small doubt..i will assign the full points, if my issue get resolved.
    Our functional consultant is asking to add POST1 field(from PRPS table) as WBS element description. Actually it is not available in the datasource, which you have mentioned. Now what i need to do.Please suggest me.
    Thanks,
    Kameswari

  • How to load data from a virtual cube with services

    Hello all,
    we have set up a virtual cube with service and create a BEx report to get the data from an external database. That works fine. The question is now:
    Is it some how possible to "load" the data from this virtual cube with service (I know that there are not really data...) into an other InfoCube?
    If that is possible, can you please give my some guidance how to set up this scenario.
    Thanks in advance
    Jürgen

    Hi:
    I don't have system before me, so try this.
    I know it works for Remote Cube.
    Right Click on the Cube and Select Generate Export Data Source.
    If you can do this successfully, then go to Source Systems tab and select the BW. Here, Right CLick on select Replicate DataSources.
    Next, go to InfoSOurces, click on Refresh. Copy the name of Virtual Cube and add 8 as a prefix and search for the infosource.
    If you can see it, that means, you can load data from this cube to anywhere you want, just like you do to ODS.
    ELSE.
    Try and see if you can create an InfoSpoke in Virtual Cube. Tran - RSBO.
    Here, you can load to a database table and then, from this table, you can create datasource, etc.
    ELSE.
    Create query and save it as CSV file and load it anywhere you want. This is more difficult.
    Good luck
    Ram Chamarthy

  • Query on Virtual cube 0FIGL_V10 show old "Last Data Updated:" on Portal

    Hello all,
    We are using the SAP delivered virtual cube for G/L financial statements reporting. We are running these queries on Portal (BI 7.0) we are using the 0ANALYSIS_PATTERN template. Now since we are not actually loading data in this cube the template show "Last Data updated" as long time back, I am guessing when the cube was last activated or transported.
    The template and time stamp is working perfect for queries on all other cubes or Multiproviders.
    The business users are asking can we have the actual time stamp there (when actually the data was last loaded),
    Does anyone have any idea how we can do it without actually rewriting the queries on standard cube.
    Thanks in advance,

    Hi Kiran,
    Always last update data status only will see in our reports on last data updated date. It may be activation of data in cube. please check is there any option of activate data in data flow level.
    Try to create one record in remote data source  and load it again. It may be find the latest status.
    Thanks,
    Chandra

Maybe you are looking for