Missing Codition Types at 2LIS_13_VDKON data source  in RSA3 Tcode.

Dear All,
i am getting problam with 2LIS_13_VDKON data source in RSA3 Tcode.
When i am checking it, some condition types are missing.
i checked it condition types are active also.
and reloaded setup table.
Still same problam is coming .
please provide me your inputs.
Regards
SSMS

hello
How did you solve your problem. I m getting exactly same issue where some cond types (active) are missing.

Similar Messages

  • Missing Connection Types in Report Data Source Properties

    I am running three installations of SharePoint 2013 with SQL 2012 for Business Intelligence, with Kerberos working.
    I have been having intermittent issues with SQL Reporting Services in a multi-server configuration, but finally had it working in all three environments.  We discovered many of our issues were caused by master page customizations which we have now disabled
    temporarily while we work on function issues.
    However, we started having a new problem with Reporting Services on just ONE of the environments, and we are not sure what happened.  Most everything appears to still be working, except Reporting Services. I have verified that features are enabled
    in both the farm and the site collection, as well as insured managed content types are being allowed on the target document libraries.  I also verified the Reporting Service properties and database connections.  However, I have found two clear symptoms
    which I can find no other information regarding (in logs or in online searches):
    1) If we try to create a new "Report Data Source" using the SharePoint UI, it will open a new Data Source Properties window, but in the "Data Source Type" field, there are no values available for me to select my connection type (e.g.,
    Microsoft SQL Server).
    2) In Central Administration, while administering the Reporting Services Service Application, I can manage a majority of the settings, including System Settings, Manage Jobs, Key Management, etc.  However, if I click on the link for "E-mail Settings,"
    I get the dreaded "Sorry Something went wrong screen," with an error saying that "Object reference is not set to an instance of an object."
    So I am hoping someone has an idea of where I can look for additional information on this issue, or even better, knows what I am missing to get Reporting Services functional again in this environment!
    Thanks in advance for any help provided.
    Elohir

    However, we started having a new problem with Reporting Services on just ONE of the environments, and we are not sure what happened.
    Hi Elohir,
    I assume you are installing Reporting Services SharePoint mode for SharePoint 2013, you need to:
    1. Install "Reporting Services – SharePoint" and "Reporting Services add-in for SharePoint Products" on the SharePoint server that the SSRS service will run on.  All other SP servers only need the "Reporting Services
    add-in for SharePoint Products."   
    2. Run the Insall-SPRSService and Install-SPRSServiceProxy on Central Admin and SSRS server.
    3. Start SSRS Service only on desired SSRS servers.
    http://msdn.microsoft.com/en-us/library/jj219068.aspx
    Since the issue occurs only to one of the environments, please provide more information about your multi-server configuration, and point out the server where the issue occurs.
    Regards,
    Please remember to mark the replies as answers if they help and unmark them if they provide no help. If you have feedback for TechNet Subscriber Support, contact
    [email protected] .
    Rebecca Tu
    TechNet Community Support

  • Power View in SharePoint Server - The data extension type for a data source is not valid

    Hi All,
    All of a sudden I am getting following error when trying to create Power View report using shared report data source (no error testing the connection):
    "The current action cannot be completed. The data extension type for a data source
    'http://dev/Shared Ducuments/Sales.rsds' is not valid for this operation"
    I already have a data source (I had created it after creating my site collection a week ago) and when I use this source to create Power View report then there is no error but I am getting above error when I create another similar data source and use it to create
    a Power View report.
    Please help me to resolve the error.
    Thanks

    I am going nuts! I had selected 'Analysis Services' instead of 'Microsoft BI Semantic Model for Power View'

  • Bringing the actual cost (splitting among activity type) using 0co_om_cca data sources

    Hi expert,
    I'm using the data source 0co_om_cca_9 to bring the cost center data but the actual data is not being split per fixed and variable cost.
    I've checked with CO team and the process to split the cost happens in the end of the month and it record the data with value type '3' in the covp table but the standard data source 0co_om_cca_9 just get the data with value type '4'.
    Does someone know a standard datasource to bring the data with value type 3 from covp?
    Thank you.
    César Sevilha.

    Hi César,
    Please check if DataSource 0CO_OM_CCA_2 can deliver what you need.
    Best regards,
    Sander

  • KSCHL (condition type) for Purchasing data source?

    hi experts,
    here is my client requirement.
    The Purchasing department needs the possibility to make analysis on Purchasing data. E.g. to support the purchasing department in negotiations with the vendors. They must be able make analysis on
    material,
    vendor,
    quantity invoiced,
    quantity received,
    values (price) in both document currency and local currency.
    The bust be able the see the price  u201CPB00 Gross Priceu201D and all Pricing Elements to the be able to see the real cost of the purchased material e.g. u201CFRC1 Freight/Quantityu201D
    through which data source i can get the Quantities and values, and from which table i can get the KSCHL so that i can enhance to my data source.
    in sdn i have seen KSCHL from KONV table. if that is the table can you plz specify the data source and logic to meet my requirements.
    regards
    venuscm
    Edited by: venugopal vadlamudi on Nov 11, 2010 5:52 AM
    Edited by: venugopal vadlamudi on Nov 11, 2010 10:30 AM

    Hi Venugopal,
    The Datasource that you need to enhance would definitely be 2LIS_02_ITM.
    To fetch the gross price you can use the below mentioned logic:
    DATA: l_t_MC02M_0ITM LIKE MC02M_0ITM OCCURS 0 WITH HEADER LINE,
          wa_MC02M_0ITM LIKE LINE OF l_t_MC02M_0ITM,
          tp_KBETR LIKE KONV-KBETR,
          tp_KPOSN LIKE KONV-KPOSN,
          tp_KNUMV LIKE EKKO-KNUMV.
    LOOP AT C_T_DATA INTO wa_MC02M_0ITM.
          CLEAR tp_KBETR.
          CLEAR tp_KPOSN.
          CLEAR tp_KNUMV.
          tp_KPOSN = wa_MC02M_0ITM-EBELP.
          SELECT SINGLE
          KNUMV
          INTO tp_KNUMV
          FROM EKKO
          WHERE EBELN EQ wa_MC02M_0ITM-EBELN.
    *     Select query to Fetch the Gross Price (KBETR) for a PO
          SELECT SINGLE
          KBETR
          INTO tp_KBETR
          FROM KONV
          WHERE KNUMV EQ tp_KNUMV AND
                KPOSN EQ tp_KPOSN AND
                KSCHL LIKE 'PB%'.
          IF SY-SUBRC EQ 0.
            wa_MC02M_0ITM-YYKBETR = tp_KBETR.
            MODIFY C_T_DATA FROM wa_MC02M_0ITM.
          ENDIF.
      ENDLOOP.
    You may still want to check with the functional person for the exact logic.
    Regards,
    Hemant Khemani

  • Short Dump Error when we check the Data Source in RSA3

    Hi. Experts.
                       We are using  Data Source 2Lis_13_VDITM and when we try to check in RSA3
    for recorts it is going short dump. I am just providing the error information.
                  An exception occurred that is explained in detail below.
    The exception, which is assigned to class 'CX_SY_IMPORT_MISMATCH_ERROR', was
    not caught in procedure "MCEX_BW_LO_API" "(FUNCTION)", nor was it propagated by a RAISING
    clause.
                   Since the caller of the procedure could not have anticipated that the exception would occur, the current program is terminated.
    The reason for the exception is:
    When importing the object "MC13VD0ITM_TAB", the component no. 96 in the dataset has a different type from the corresponding component of the target object in the program "SAPLMCEX".
                        The data type is "N" in the dataset, but "X" in the program.
            Since I don't have much knowledge in ABAP, Please help me out regarding this issue. Suggest me possible solution and let me know if you need any further information.

    Hi,
    Try adding a TRY CATCH block in the extractor function module to catch the exception.
    For the type mismatch you will have to make sure the variables are of same type or they can be type converted.
    Use the debug mode of RSA3 to find what data is being exchanged.
    You will need to change your extractor to avoid the dump.. or ensure that the value being passed to that object is correctly being passed(correct format)
    Cheers
    Kiran

  • Errors occurred during the extraction for CFM data sources in RSA3

    Hi Friends,
                   When i try to execute the datasource's in Extractor Checker (RSA3) using the datasources given below:
    0CFM_DELTA_POSITIONS
    0CFM_INIT_POSITIONS
             I'm getting the error message saying that "Errors occurred during the extraction"
              This is my frist error and the the second error what i'm having is:
    When i try to execute the the datasources given below:
    0CFM_RDB_AN_1
    0CFM_RDB_AN_2
    0CFM_RDB_AN_3
    0CFM_RDB_AN_4
    0CFM_RDB_AN_5
    0CFM_RDB_AN_6
             It is asking "Portfolio Hierarchy" in RSA3 selections, even though i've activated portfolio hierarchy, i can't able to get that hierarchy in Extractor Checker Selections (RSA3), unless i get that hierarchy, i can't be able to excute any datasource in extractor checker.
    Awaiting for your inputs with curiousity.
    Thanx in advance,
    BalajiReddy

    Have a look at the following OSS Notes :
    1029605
    808387
    995660
    594272

  • ZPRS- Cost(data source H) condition type is not visible in CRM sales order

    Hi Colleagues,
    I created a sales order in ERP. In the sales order the condition type ZPRS- Cost(data source H) is available. I downloaded the this sales order to CRM i can able to see the ZPRS condition type in sales order.
    But when i create a new sales order with the same data (BP, Material,Sales organization) the condition ZPRS is not visible in the sales order. When i checked the PRC_TRACE it is showing the error as "Condition with external datasource has not be found".
    Can any body help me in this issue.
    Tanks and Best Regards,
    Rajeswari

    Hi Michael ,
    Thank you very much for your answer. I just require one more confirmation from you.
    Could you please confirm If we post a returns document in CRM system directly the "ZPRS or VPRS - Cost" Condition type will not appear in the return document condition types.
    Untill and unless it is a billing document in CRM system the "cost" condtion wont appear rt?
    Could you please confirm it.
    Thanks in advance for your help.
    Best Regards,
    Rajeswari

  • Data source on top of virtual cube s_r_infoprov  expection

    hi experts,
    i have generated a export datasource on virtual cube. The virtual cube is based on a function module.
    When I display data in virtual cube it shows all the data .When I extract the data from it through DTP , it works well.
    But when i am using data source(exported data source) to pull the data from the virtual cube it is not pulling the data.
    It is throwing an exception at the following place.
    CALL FUNCTION 'RSDRI_INFOPROV_READ'
           EXPORTING
                                                            "begin qyh030513
             i_infoprov             = l_infoprov  "'/CPMB/ASFIN'
    *        i_infoprov             = do_util->do_appl_info_m-MULTIPROV  "'/CPMB/ASFIN'
                                                              "end qyh030513
             i_th_sfc               = lth_sfc
             i_th_sfk               = lth_sfk
             i_t_range              = lt_range
             i_rollup_only          = space
             "i_use_db_aggregation   = abap_true "abap_true "RS_C_TRUE
             i_use_db_aggregation   = IF_DB_AGGREGATE "abap_true "RS_C_TRUE
             i_use_aggregates       = abap_true "abap_true "RS_C_TRUE
             i_packagesize          = i_packagesize
             i_authority_check      = space
             i_currency_conversion  = space
           IMPORTING
             e_t_data               = et_data
             e_split_occurred       = e_split_occurred
             e_end_of_data          = e_end_of_data
             e_stepuid              = l_stepuid
           CHANGING
             c_first_call           = c_first_call
           EXCEPTIONS
             illegal_input          = 1
             illegal_input_sfc      = 2
             illegal_input_sfk      = 3
             illegal_input_range    = 4
             illegal_input_tablesel = 5
             no_authorization       = 6
             illegal_download       = 8
             illegal_tablename      = 9
             OTHERS                 = 11.
    When we go inside this function module in debugging the below code will be there
    STATICS: s_r_infoprov   TYPE REF TO cl_rsdri_infoprov.
       CLEAR: e_t_msg.
       IF i_clear = rs_c_true.
         CLEAR s_r_infoprov.
         RETURN.
       ENDIF.
    IF s_r_infoprov IS NOT INITIAL AND c_first_call = rs_c_true.
    *   nested call of RSDRI_INFOPROV_READ
         MESSAGE e882(dbman) RAISING illegal_input.
       ENDIF.
    Problem is with the s_r_infoprov value.  When I execut with data source in RSA3 then  the value is {O:222*\CLASS=CL_RSDRI_INFOPROV}
    and when I executed directly the virtual cube the value for s_r_infoprov is initial.
    So because of this value {O:222*\CLASS=CL_RSDRI_INFOPROV} the following code got executed and throwing an exception.
    Experts please guide how to solve this. will data source on top of virtual cube wont work?
    thank you

    Hi Sander,
    Thank you for the reply.
    I am trying to understand why an virtual cube is not working if it is generated as an exported  data source and it is working perfectly if directly connected as a source to DSO or Cube?
    Problem is with the s_r_infoprov value. It is getting populated differently , that is why there is an exception which using data source(virtual cube generated as exported data source).
    If this functionality is not going to work then why SAP has given an option to generate an exported DS of virtual cube.
    thank you for all giving some valuable inputs on this. Some how some thing is missing it seems....

  • Regarding: Loading data from R/3 To BI for a Generic Data source

    Hi Every,
    Need Help Urgent
    I had created a Generic data source with function Module as the data source, in Rsa3 it is working fine.
    1-> I had replicated the data source to Bi then i had created a info package and then I executed the same. when it is getting the data and it is show in the request monitor (Number of records )  but the status is not changing from Yellow to Green.
    Status in Step by step analysis is every step is green Except "Data selection successfully finished ?"  (RED)
    2-> Then I had seen the Back ground job in Source system which is executing still. I waited for it for a long time but nearly 30 min.
    (I had done Steps one and 2 number of time by activating replicating the data source and so on, but still their is no change in it)
    3-> Then I had canceled that back ground job with the help of BASIS (as i feel that it is something going wrong).
    4-> I feel that their is some thing wrong in the Code of Extractor.
    Please Help...............
    ""Local Interface:
    *"  IMPORTING
    *"     VALUE(I_REQUNR) TYPE  SRSC_S_IF_SIMPLE-REQUNR
    *"     VALUE(I_DSOURCE) TYPE  SRSC_S_IF_SIMPLE-DSOURCE OPTIONAL
    *"     VALUE(I_MAXSIZE) TYPE  SRSC_S_IF_SIMPLE-MAXSIZE OPTIONAL
    *"     VALUE(I_INITFLAG) TYPE  SRSC_S_IF_SIMPLE-INITFLAG OPTIONAL
    *"     VALUE(I_READ_ONLY) TYPE  SRSC_S_IF_SIMPLE-READONLY OPTIONAL
    *"     VALUE(I_REMOTE_CALL) TYPE  SBIWA_FLAG DEFAULT SBIWA_C_FLAG_OFF
    *"  TABLES
    *"      I_T_SELECT TYPE  SRSC_S_IF_SIMPLE-T_SELECT OPTIONAL
    *"      I_T_FIELDS TYPE  SRSC_S_IF_SIMPLE-T_FIELDS OPTIONAL
    *"      E_T_DATA STRUCTURE  ZBI_MATGRIR OPTIONAL
    *Need to get the data only for two Gl account which are fro material purchase while MIGO
    *G/L Account Numbers: 0010502001 0010502002
    data: E_T_DATA1 type table of ZBI_MATGRIR.
      RANGES: R_BUKRS FOR BSIS-BUKRS,
              R_BUDAT FOR BSIS-BUDAT,
              R_GJAHR FOR BSIS-GJAHR,
              R_HKONT FOR BSIS-HKONT.
      DATA: L_S_SELECT TYPE SRSC_S_SELECT.
      STATICS: S_S_IF TYPE SRSC_S_IF_SIMPLE,
               S_COUNTER_DATAPAKID LIKE SY-TABIX,
               S_CURSOR TYPE CURSOR.
    *Declare
      TYPES: BEGIN OF TY_FAGL,
        RBURS TYPE FAGLFLEXA-RBUKRS,
        RYEAR TYPE FAGLFLEXA-RYEAR,
        DOCNR TYPE FAGLFLEXA-DOCNR,
        BUZEI TYPE FAGLFLEXA-BUZEI,
        DOCLN TYPE FAGLFLEXA-DOCLN,
        PRCTR TYPE FAGLFLEXA-PRCTR,
        SEGMENT TYPE FAGLFLEXA-SEGMENT,
      END OF TY_FAGL.
      DATA: GT_FAGL TYPE TABLE OF TY_FAGL,
            GS_FAGL TYPE TY_FAGL.
      IF I_INITFLAG = SBIWA_C_FLAG_ON.
        CASE I_DSOURCE.
          WHEN 'ZFI_GL_M4'.
          WHEN OTHERS.
            IF 1 = 2. MESSAGE E009(R3). ENDIF.
         this is a typical log call. Please write every error message like this
            LOG_WRITE 'E'                  "message type
                      'R3'                 "message class
                      '009'                "message number
                      I_DSOURCE   "message variable 1
                      ' '.                 "message variable 2
            RAISE ERROR_PASSED_TO_MESS_HANDLER.
        ENDCASE.
        APPEND LINES OF I_T_SELECT TO S_S_IF-T_SELECT.
    Fill parameter buffer for data extraction calls
        S_S_IF-REQUNR    = I_REQUNR.
        S_S_IF-DSOURCE = I_DSOURCE.
        S_S_IF-MAXSIZE   = I_MAXSIZE.
        APPEND LINES OF I_T_FIELDS TO S_S_IF-T_FIELDS.
      ELSE.
    Data transfer: First Call      OPEN CURSOR + FETCH
                   Following Calls FETCH only
    First data package -> OPEN CURSOR
        IF S_COUNTER_DATAPAKID = 0.
    Fill range tables BW will only pass down simple selection criteria
    of the type SIGN = 'I' and OPTION = 'EQ' or OPTION = 'BT'.
          LOOP AT S_S_IF-T_SELECT INTO L_S_SELECT WHERE FIELDNM = 'BUKRS'.
            MOVE-CORRESPONDING L_S_SELECT TO R_BUKRS.
            APPEND R_BUKRS.
          ENDLOOP.
          LOOP AT S_S_IF-T_SELECT INTO L_S_SELECT WHERE FIELDNM = 'GJAHR'.
            MOVE-CORRESPONDING L_S_SELECT TO R_GJAHR.
            APPEND R_GJAHR.
          ENDLOOP.
          LOOP AT S_S_IF-T_SELECT INTO L_S_SELECT WHERE FIELDNM = 'BUDAT'.
            MOVE-CORRESPONDING L_S_SELECT TO R_BUDAT.
            APPEND R_BUDAT.
          ENDLOOP.
    *GRIR Inventory (RM/Stores/Spares/FG)   10502001
    *GRIR Services & Others Payable   10502002
          R_HKONT-SIGN = 'I'. "i_t_select-sign.
          R_HKONT-OPTION = 'BT'." i_t_select-option.
          R_HKONT-LOW = '0010502001'.
          R_HKONT-HIGH = '0010502002'. "i_t_select-high.
          APPEND R_HKONT.
    Determine number of database records to be read per FETCH statement
    from input parameter I_MAXSIZE. If there is a one to one relation
    between DataSource table lines and database entries, this is trivial.
    In other cases, it may be impossible and some estimated value has to
    be determined.
          OPEN CURSOR WITH HOLD S_CURSOR FOR
           SELECT BUKRS
                  AUGBL
                  ZUONR
                  BELNR
                  GJAHR
                  BUZEI
                  BUDAT
                  HKONT
                  BLART
                  MONAT
                  BSCHL
                  SHKZG
                  DMBTR
                  WAERS
                  FROM BSIS
                 INTO TABLE E_T_DATA
                  WHERE BUKRS  IN R_BUKRS
                    AND GJAHR IN R_GJAHR
                    AND BUDAT IN R_BUDAT
                    AND HKONT IN R_HKONT.
    Fetch records into interface table.
      named E_T_'Name of extract structure'.
        FETCH NEXT CURSOR S_CURSOR
                   APPENDING CORRESPONDING FIELDS
                   OF TABLE E_T_DATA1
                   PACKAGE SIZE S_S_IF-MAXSIZE.
        IF SY-SUBRC <> 0.
          CLOSE CURSOR S_CURSOR.
          RAISE NO_MORE_DATA.
        ENDIF.
    DELETE E_T_DATA WHERE BLART NE 'WE'.
          SELECT BUKRS
                 AUGBL
                 ZUONR
                 BELNR
                 GJAHR
                 BUZEI
                 BUDAT
                 HKONT
                 BLART
                 MONAT
                 BSCHL
                 SHKZG
                 DMBTR
                 WAERS
                 FROM BSAS
                 into table   E_T_DATA
                 WHERE BUKRS  IN R_BUKRS
                   AND GJAHR IN R_GJAHR
                   AND BUDAT IN R_BUDAT
                   AND HKONT IN R_HKONT.
               FETCH NEXT CURSOR S_CURSOR
                  APPENDING CORRESPONDING FIELDS
                  OF TABLE E_T_DATA
                  PACKAGE SIZE S_S_IF-MAXSIZE.
    append LINES OF e_t_data1 TO E_T_DATA.
    DELETE E_T_DATA WHERE BLART NE 'WE'.
    ENDIF.                             "First data package ?
        DATA: F_YEAR TYPE BKPF-GJAHR.
        DATA: F_PERI TYPE BAPI0002_4-FISCAL_PERIOD.
    IF E_T_DATA[] IS NOT INITIAL.
       SELECT RBUKRS
              RYEAR
              DOCNR
              BUZEI
              DOCLN
              PRCTR
              SEGMENT
              FROM FAGLFLEXA
              INTO TABLE GT_FAGL
              FOR ALL ENTRIES IN E_T_DATA
           WHERE RYEAR = E_T_DATA-GJAHR
               AND DOCNR = E_T_DATA-BELNR
               AND RLDNR = '0L'
               AND RBUKRS = E_T_DATA-BUKRS
               AND BUZEI = E_T_DATA-BUZEI.
             WHERE RYEAR = E_T_DATA-GJAHR
               AND DOCNR = E_T_DATA-BELNR
               AND RBUKRS = E_T_DATA-BUKRS.
         AND DOCLN = E_T_DATA-BUZEI.
    ENDIF.
    LOOP AT E_T_DATA.
          IF E_T_DATA-SHKZG = 'H'.
            E_T_DATA-DMBTR = E_T_DATA-DMBTR * -1.
          ENDIF.
          CLEAR: F_YEAR.
          CALL FUNCTION 'BAPI_COMPANYCODE_GET_PERIOD'
            EXPORTING
              COMPANYCODEID       = E_T_DATA-BUKRS
              POSTING_DATE        = E_T_DATA-BUDAT
           IMPORTING
             FISCAL_YEAR         = F_YEAR
             FISCAL_PERIOD       = F_PERI.
          DATA: V_DOC(6) TYPE C .
          CLEAR: V_DOC.
          V_DOC =  E_T_DATA-BUZEI.
          IF V_DOC  IS NOT INITIAL.
            CALL FUNCTION 'CONVERSION_EXIT_ALPHA_INPUT'
              EXPORTING
                INPUT  = V_DOC
              IMPORTING
                OUTPUT = V_DOC.
          ENDIF.
         aS PROFIT center is not updated in all the lines in Bsis
          READ TABLE GT_FAGL INTO GS_FAGL WITH KEY  RYEAR = E_T_DATA-GJAHR
                                                    DOCNR = E_T_DATA-BELNR
                                                    RBURS = E_T_DATA-BUKRS
                                                    BUZEI = E_T_DATA-BUZEI.
          IF SY-SUBRC = 0.
            E_T_DATA-PRCTR = GS_FAGL-PRCTR.
            E_T_DATA-SEGMENT = GS_FAGL-SEGMENT.
          ENDIF.
    *As we are using the amount DMBTR in which the amount
    *will be in company code currency that is Local currency
    *group currency always in the main company code currency.
          CONCATENATE   F_YEAR '0' F_PERI INTO E_T_DATA-FISCPER.
          MODIFY E_T_DATA. " from gs_bsis transporting dmbtr fiscper.
          CLEAR: E_T_DATA.
        ENDLOOP.
          S_COUNTER_DATAPAKID = S_COUNTER_DATAPAKID + 1.
       ENDIF.

    Hi,
    Please check the log of same jobs for last week and check is today its taking more time,also check with basis is their any backup initited at same time.
    Moreover until the background job failed automatically its difficult to imagine what is the exact issue.
    Thanks,

  • Report using excel sheet as data source not working when deployed to production.

    I have a set of SSRS reports that use an excel sheet as a datasource.
    Now when I preview the reports or run them in the Visual Studio IDE on my local machine i.e
    the development machine, the reports run fine & display all correct results.
    However, when we deploy the reports to the Production machine which is on a separate server,
    the following error message is displayed :
    "The current action cannot be completed. The user data source credentials do not meet the requirements to run this report. Either the user data source credentials are not stored in the report server database, or the user data source is configured not to
    require credentials but the unattended execution account is not specified. (rsInvalidDataSourceCredentialSetting)"
    I've copy pasted the connection string used for connecting to the excel file as reference :
    Provider=Microsoft.ACE.OLEDB.12.0;Data Source=C:\MyFolder\CurrentFile.xlsx;Extended Properties="Excel 12.0;HDR=Yes;";
    The excel file has also been uploaded to the machine housing the reporting server (i.e to C:\MyFolder location)

    Hi Aaakar,
    From the error message, please try to change the setting of Credential type for the data source on report manager as follows:
    Go to the Data Sources properties page.
    For the Connect Using option, select Credentials stored securely in the report server.
    In User Name and Password, type credentials that can be used to access the database. If you are using SQL Server as the data source, the user name must be valid for both logging on to the server and for accessing the database that contains the data for
    the report.
    If the user name and password are credentials for a Windows account, select Use as Windows Credentials.
    If you want the report server to pass the credentials of the user accessing the report to the server hosting the external data source, click Windows Integrated Security. In this case, the user is not prompted to type a user name or password.
    For more details about Configure Data Source Properties, please see:
    http://technet.microsoft.com/en-us/library/ms155882.aspx
    Hope this helps.
    Thanks,
    Katherine Xiong
    Katherine Xiong
    TechNet Community Support

  • Key Figures in Joined Data Source

    Hello,
    I have created some Key Figures for the Data Source "Sales Order Item". Reports based on this Data Source are working fine and also my Key Figures are displaying the right results.
    Now I want to create a Joined Data Source (Left Outer Join) with the Sales Order Item and Outbound Delivery Detail Line Items. The problem I have here is, that my own Key Figures based on Sales Order Item are not showing up in the list and I cannot select them.
    Does anyone know why? Do I have to create the same Key Figures again and again as soon as I want to join the Sales Order Item Data Source with any other Data Source?
    Thanks for any feedback.

    Key figure are not available in joining Data Sources, Characteristics of same type from different Data Source data element available in joining scenario.
    Regards
    Naveen

  • Dynamic File Name in Data Source?

    In Data Model, there is a option "File" as Type. My question is can file name be dynamic, say I set a parameter "ReportID", can this parameter be part of file name. something like
    {$ReportID}.xml
    or
    {:ReportID}.xml
    If yes, what is the syntax.
    Thanks.

    I don't think you can do that. When you speficy 'File' as the type for the Data source, it expects an existing XML data file stored in the directory that has been set up by the Administrator.
    There is no way to define a parameter for the file name and then use this instead..
    Thanks!
    Edited by: BIPuser on Mar 1, 2010 1:10 PM

  • Sessions.xml data source not working using toplink

    I'm having some difficulty getting a data source to work. I have an ADF selectOneChoice component and I want the list of values available to be bound to a database given by the JNDI name and controlled at the j2ee container level.
    I have created a sessions.xml file
    <toplink-configuration>
    <session>
    <name>PIMSSession</name>
    <project-xml>META-INF/PIMS_ADF_TLMap.xml</project-xml>
    <session-type>
    <server-session/>
    </session-type>
    <login>
    <datasource>jdbc/PIMSDS</datasource>
    <uses-external-connection-pool>true</uses-external-connection-pool>
    </login>
    <connection-pool>
    <is-read-connection-pool>true</is-read-connection-pool>
    <name>default</name>
    <login>
    <datasource>jdbc/PIMSDS</datasource>
    <platform-class>Oracle10g</platform-class>
    <uses-external-connection-pool>true</uses-external-connection-pool>
    </login>
    </connection-pool>
    </session>
    </toplink-configuration>
    The sessions.xml points to the toplink mapping file PIMS_ADF_TLMap.xml, which contains information about the database view that is used to populate the drop down list. It also contains a deployment connection:
    <toplink:login xsi:type="toplink:database-login">
    <toplink:platform-class>oracle.toplink.platform.database.oracle.Oracle10Platform</toplink:platform-class>
    <toplink:user-name>pimsdev</toplink:user-name>
    <toplink:password>BB742416276274A494F7008B3AE0FB10</toplink:password>
    <toplink:driver-class>oracle.jdbc.OracleDriver</toplink:driver-class>
    <toplink:connection-url>jdbc:oracle:thin:@yvworad01:1521:PIMSD</toplink:connection-url>
    </toplink:login>
    The problem I'm having is that my ADF application always binds the selectOneChoice to the connection-url in the toplink mapping file instead of the datasource in the sessions.xml.
    If I remove the deployment connection from the mwp file, recompile and deploy I get a JBO-29000 error about not having a suitable driver.
    Does anyone know how to get the selectOneChoice to bind to a datasource instead of the toplink mapping connection-url?

    Hi Aaakar,
    From the error message, please try to change the setting of Credential type for the data source on report manager as follows:
    Go to the Data Sources properties page.
    For the Connect Using option, select Credentials stored securely in the report server.
    In User Name and Password, type credentials that can be used to access the database. If you are using SQL Server as the data source, the user name must be valid for both logging on to the server and for accessing the database that contains the data for
    the report.
    If the user name and password are credentials for a Windows account, select Use as Windows Credentials.
    If you want the report server to pass the credentials of the user accessing the report to the server hosting the external data source, click Windows Integrated Security. In this case, the user is not prompted to type a user name or password.
    For more details about Configure Data Source Properties, please see:
    http://technet.microsoft.com/en-us/library/ms155882.aspx
    Hope this helps.
    Thanks,
    Katherine Xiong
    Katherine Xiong
    TechNet Community Support

  • Error while loading data into PSA from data source

    Hi Experts ,
    I am loading purchase order history data from standerd datasource 0SRM_TD_CF into PSA using infopackage.However,whenevr I run infopackage,the call monitor shows the status as yellow.I monitored the process for half an hour but it never ended.I have verified following things from my side.
    1.Checked whether datasource is properly replicated into BI system.
    2.Checked if source system contains data.
    3.Since i am loading transaction data,I have kept update as "initialize delta process"
    Please let me know whether I am missing anything.
    Regards,
    Mandar.

    Hi mandarha,
    Do the following steps
    1)Check the job log in source systems.
    2)Check the Data Source in RSA3(Extracter checker) on Source systems side.give DS name press enter then execute,If there is no issue with DS,it will give pop-up with records.If it has any errors first rectify it and then load it again.
    3)In details tab of IP,check whether extraction completed or not.
    4)Are you running the IP with option "Intialse with data trasfer" or "Without data transfer"
    If you running with data transfer and without data selection,usually it will take time based on the volume of historical data.(to find how many records it picks do the 2 step)
    Hope this helps
    Regards,
    Venkatesh.

Maybe you are looking for