SPM: Using "refresh data sources" in Data Source Administrator

Hi All,
When making changes to a query, that I want to be reflected in the flex front end, I carry out the following steps:
1) save the query in query designer
2) load up the SPM user interface, navigate to Administration -> Data Source Administration
3) Click on "Refresh All Data Sources"
4) Wait for the "wait" cursor to finish (only takes a few seconds)
5)log out of the portal
6) clear the browsers cache
7) reload the web page
8) Repeat Steps 2-7 again
9) the query's changes are not reflected in the UI
Does anyone know what I should be doing differently so that the query changes are reflected in the Data Source?
Kind regards,
Neil

Hi Neil,
You are almost right on the steps, just a few changes.
1. Modify your query and save the query changes.
2. Close the query designer so you are not locking the query (optional step, but recommended).
3. Go to transaction RSRT and generate the query (optional step, but recommended).
4. Login to portal and launch SPM user interface, navigate to Administration -> Data Source Administration
5. Click on "Refresh All Data Sources"
6. Check the UI logs (Settings > View logs), to confirm that datasource refresh has finished successfully (your should receive messages saying 'Metadata Loaded'.  The cursor will turn into a clock multiple times, this is the UI checking to see if the metadata refresh has finished.  You should see the message 'Metadata Loaded' as many times as there are registered BW queries.  This could take a few minutes depending on the number of queries registered in SPM UI.
You need not clear your browser cache or logout of the application.  At most you may need to relaunch the application.
These steps should be sufficient to reflect any query changes to the SPM UI.
Regards,
Rohit

Similar Messages

  • Using Web Service as Data source (Pluggable Data Source) in Oracle Reports

    Anyone using Webservice as a pluggable data source in oracle 10g reports.(Report builder version:10.1.2.0.2)? We need to be able to use web service as one of the data source to create reports.Tried the following and ran into issues:
    Imported the wspds.jar fle from oracle plugin exchage. When tried using it received the following error in oracle reports trace file.
    04/26 22:27:22 java.lang.NoSuchMethodError: oracle.xml.parser.schema.XSDBuilder.build([Loracle/xml/parser/v2/XMLDocument;Ljava/net/URL;)Ljava/lang/Object;
    By troubleshooting issue found out that xmlparserv2.jar in 10g is newer than xmlparserv2.jar in 9i. wspds.jar is using xmlparserv2.jar in 9i as it was developed and tested for 9i only. Believe there is no newer wspds.jar file developed specifically for 10g?
    Is there a better/standard solution to use webservice as data source in oracle 10g reports? Does later version of oracle reports supports webservice as a data source without any additional plugins? Any suggestion would be helpful.

    Please provide support for the above request.
    It appears that Oracle 10g reports doesn't support Web Service as pluggable data source. Is this option is available in Oracle 11g Reports or Oracle Discoverer Or OBIEE?

  • How to use Power Query to load data from a single source into multiple tables

    Hi all,
    I have a requirement to load my data into three different data models using Power Query. Is that possible?
    My source is a SharePoint survey list, with similar questions like:
    1) Course lecturer - John Doe
    1a) The course was useful (rate 1 to 5)
    1b) The lecturer displayed good knowledge of topic (rate 1 to 5)
    2) Course Lecturer - Mary Brown
    2a) The course was useful (rate 1 to 5)
    2b) The lecturer displayed good knowledge of topic (rate 1 to 5)
    I would like to split the data into separate data models (one for John Doe; another for Mary Brown), so that I can compare the different lecturers. Other than running separate surveys for each of them, I thought of using Power Query to transform the data.
    Is it possible?
    Thanks.
    Regards
    GM

    Yes, this is possible.
    Start with a single query that returns you the data for all lecturers.
    Right-click on the "all lecturers" query in the queries pane, and choose Reference, once for each lecturer. This will create a query for each lecturer.
    Open the query for each lecturer and filter the data so that only that lecturer's results are visible.
    Click "Close & Load To..." for each lecturer's query  to load the data into the data model. This will create a data model table for each lecturer.
    If your question is more about how to transform such a survey list into a table that can be easily filtered, please provide an example of how the list shows up in Power Query.
    Ehren

  • UME using SAP R/3 as Data Source

    Hi,
    We are trying to set User authentication to SAP R/3 system, not load balanced system, on the User Management Configuration values: Client=501, Userid=sapjsf, Password=pwd, sys id=RS1, Group and Message server= blank, Application server= server.company.com, Sys. number=00, Max pool=10, Max wait=300000.
    When testing connection, I get this message:
    (System ID): com.sap.mw.jco.JCO$Exception: (101) RFC_ERROR_PROGRAM: 'mshost' missing
    (System ID & System Number): OK
    Is this an error? since our SAP R/3 is not a load balanced system.
    Did we miss any item for the setup, in dataSourceConfiguration_r3.xml? The SAPJSF "communication user" got the right sap role and authorizations.
    Portal version : EP6 SR1
    Regards
    Huzaifah

    Hi,
    If u want to Use The SAP R3 System as Data source u may
    do it from config tool if u got following message.
    WARNING! You are not allowed to select dataSourceConfiguration_r3.xml as active configuration file.
    (For Portal Patch less than SP13 u must download two data source file which is attached with note - 718383
    and upload it to portal which is described in the note)
    the following are the procedure which i apply ,
    Go to System Administration -> System Configuration ->UM  Configuration
    Now Do not change Data source from Here.
    Make sure  your data source is "Database Only"
    (dataSourceConfiguration_database_only.xml)
    Now enter the following value under SAP System Tab.
    Client : - Your sap system client
    User:-  Sap user
    password: - password
    System language:- your system language
    Application server: - Host name or IP of sap system
    System Number : -  SAP instance number
    Maximum Size of Connection Pool : -  As per req.
    Maximum Wait Time in Milliseconds :- 10000
    Now, save the changes and shutdown the portal server.
    Using Config Tool change the data source. Run the following
    <drive:\> usr\sap\<sid>\JC<instance number>\J2EE\configtool\configtool.bat
    (Make sure the portal system is shutdown)
    Under Cluster Data -> Global Server Configuration -> services -> com.sap.security.core.ume.services
    Now find the key: -  ume.persistence.data_source_configuration     
    The default was : - dataSourceConfiguration_database_only.xml
    change the value to :- dataSourceConfiguration_r3.xml     
    click on set and from flie-> apply
    Now restart the portal server ur data source changer to SAP R3 System
    Regards,
    Kaushal

  • Using COGNOS as an infoprovider/data source for WebI...?

    I saw a job posting today that listed a requirement that I am pretty sure is IMPOSSIBLE...
    WebIntelligence report experience with using COGNOS as an infoprovider/data source for BOBJ is an important skill.
    Data sources for client will be both COGNOS and SAP.
    Now, it wouldn't be the first time that a client spec asked for the IMPOSSIBLE- but it got me to wondering if maybe I am wrong in my understanding that COGNOS 8 still uses PowerPlay cubes that are 100% proprietary and can only be read with the CONGOS report-design tools.
    Is there actually a "supported" way to use COGNOS as an infoprovider/data source for BOBJ...?
    Maybe using some COGNOS equivalent of QAAS...?
    I have no intention of applying for this job - but would like to know if it is technically possible.

    It's been more than a week since I posted this question and nobody came up with a "YES" here or on the BOB Forums...so I am going to close this question as an IMPOSSIBLE request.
    Based on the current technology - COGNOS cubes can not be used as an infoprovider/data source for WebIntelligence.

  • Handling parameters when using BI Server as a Data Source

    Hi,
    Following an upgrade of our BI Publisher 10g content to 11g I've come across a syntax issue with 11g when it comes to the handling of parameters in BI Server Data Sets.
    Under 10g, the documentation stated the following:
    +"If your data source is the Oracle BI Server, use the following macro to handle the null:+
    +{$ if ${sYear}='*'$}+
    +{$elsif ${sYear}='2000' $}+
    +where Year = :sYear+
    +{$else $}+
    +where Year = :sYear+
    +{$endif$} "+
    http://docs.oracle.com/cd/E10415_01/doc/bi.1013/e12187/T518230T518233.htm
    Under 11g it does not seem to allow the use of macros defined in Data Sets and raises the following error when saving the query:
    +"java.io.IOException: prepare query failed[nQSError: 43113] Message returned from OBIS. [nQSError: 27002] Near <$>: Syntax error [nQSError: 26012] ."+
    If macros are no longer supported, what is the best approach to handle null value parameters (caused by all choices / nothing being selected)
    as they need to be defined in the Data Set WHERE CLAUSE?
    I've tried a rather clunky CASE WHEN statement which handles string parameters fine, but dates are another thing as null values are returned as 'null' text which causes TYPE mismatch errors.
    Hopefully there is an elegant way to handle this in 11g (the 10g macros were very straight forward when dealing with BI Server)
    Many Thanks

    If what you are saying is true then I would highly recommend open a ticket with Oracle Support.
    They should provide and alternative for handling NULLS in 11g.
    regards
    Jorge

  • I had no idea Aurora used the same actual App Data file as it's own source, I thought it was just emulating Firefox. I unistalled it and lost everything...

    I had installed Aurora 21.0a2 some time last year and had only used it once. I recently uninstalled it and without knowledge that it used the same exact App Data files as Firefox does, so I removed all data. Now I have lost 100's of bookmarks... I do backup my system often but not my browser data, I had only once made a copy of my Firefox data in August 2013, so I was able to copy and paste that info, which had maybe 35-40% of my bookmarks...
    I had tried Windows Recovery a couple times, it says it wasn't able to restore the data successfully each time. I had also tried Recuva, not really sure if it found the file I was searching for, so I would kindly ask any of the Forum Mods or users who would suggest this method, what "file" exactly should I search for.
    I had been searching for ".default" and ".defaults" files, and any other data relating to Firefox but with no success.
    Do I really have any options of recovery or ....?

    You should never choose to remove "personal data" when you uninstall your current Firefox version, because this will remove all profile folders and you lose personal data like bookmarks and passwords including data in profiles created by other Firefox versions.
    You will have to try to recover a recent JSON backup (bookmarks-####-##-##_xxxx.json) in the bookmarkbackups folder of that removed Firefox profile folder with Recuva or another undelete utility.
    *http://kb.mozillazine.org/Profile_folder_-_Firefox
    *https://support.mozilla.org/kb/Recovering+important+data+from+an+old+profile
    *http://kb.mozillazine.org/Transferring_data_to_a_new_profile_-_Firefox
    *http://kb.mozillazine.org/Backing_up_and_restoring_bookmarks_-_Firefox

  • How to make BI Info obj Data element in-herit source DE documntn?

    Say I am in ECC. I go to SE11, give table name MARA and then doubleclick on the dataelement 'MATNR'. Then I click on documentation. I get a popup with Short text " Material Number" and Definition "Alphanumeric key uniquely identifying the material'.
    I am interested in the latter information that is 'Definition' - & in general whatever documentation comes up for data element in ECC.CRM.SCM,SRM.
    Now I log into SAP BI. I find that under characteristic 0Material, the data element is /BI0/oimaterial. When I double-click this datamaterial and press the documentation button, the system says 'No documentation'.
    My Questions:
    1. IS there a switch we could turn on in source ECC/SRM/CRM/SRM so that the data element in SAP BI inherits the original source field data element documentation.
    { I am not too convinced of the argument- that in BI we have info objects in ECC we have fields - since I am talking of data element level information- I would tend to think of this as an oversight of the designers or de-prioritization !!}.
    2. Could we have an ABAP workaround? That is, in BI we identify the tables that house the mapping between the source and destination data elementsa and take out this information. Then we extract the dataelement documentation by function DOCU_GET (from eCC) and use the mapping info above to link SAP BI data element with source data element documentation.
    WHY do I want to punish myself as above? My use case is, we take out SAP BI Table, field, metadata etc and create a model in a modeling tool and physical implementation in a 3rd party DW database as our own canonical, application-agnostic corporate datawarehouse. I would want the source data element documentation to flow to this last system as well as the modeling tool.
    Regards
    Sasanka

    That is, in BI we identify the tables that house the mapping between the source and destination data elementsa and take out this information. Then we extract the dataelement documentation by function DOCU_GET (from eCC) and use the mapping info above to link SAP BI data element with source data element documentation.
    1) SAP don't supply this, I would imagine, because R/3 isn't the only possible source of data.  I'm currently working on an app that extracts from R/3, an Oracle database and APO.  From whence should I take the documentation for my MATERIAL info object?  While being able to transfer the documentation of data elements might be very useful for your app, I can't see that generally it would be of interest to clients - I've certainly never heard of such a requirement.  So, my opinion at least, it isn't a design flaw.
    2) As you've pointed out, you can get the tables that do the mapping, so you know the source data elements, so you can get the documentation.  I'm not sure of how to store the documentation, but the obvious candidate for a link between infoobject and dataelement would be master data of an own infoobject.  You could wrap DOCU_GET in an RFC (if it isn't RFC enabled), and do a direct call between your 3rd party app and r/3 to get the documentation.  For information about the mapping tables, I'd suggest asking that question specifically in one of the BI forums.
    matt

  • There is no source data for this data record, Message FZ205

    Hi Experts,
    I am facing a problem with the DME File download. This problem happened all of sudden in our production system since last month and it was never before. Our system landscape has also not been changed but as per our basis consultant he has added two-three more new application server to the Production client. Even we do not have this problem in our testing clients.
    Please note that we have been using the output medium '1' from the day one and thus the system has been generating the DME in 'File System' which we download on the desktop and upload the same to the bank online. After running the payment run when we trying to download the DME File, the system gives the error "There is no source data for this data record, Message FZ205".
    I tried to fix this issue through many ways but not able to. So can you please let me know the reason of this error and solution to fix this.
    With best regards,
    BABA

    Hi Shailesh,
    Please share how you solved this problem.
    Many Thanks,
    Lakshmi

  • How to get date difference in data sources

    Hi,
    In SQL:SELECT DATEDIFF(day, '2005-12-30 23:59:59.9999999', '2005-12-30 00:00:00.0000000');
    output :1
    As similar way how get the datediffence in SAP business by design of the datasources.
    Regards,
    Prasad

    Hi Prasad,
         For this you have to create a Custom DATA SOURCE for the required BO. Then Select all the required fields that you want to use it in the report for that datasource.
    The 3rd step of Data Source Wizard is shown in the image, here you can add a extra field that can hold the difference between the DateTime elements present in the BO.
    In the left column BUSINESS OBJECT ELEMENT choose DATE element (1st Date).
    In the right column TRANSFORMATION choose DateTime To Duration row and in the PARAMETER Column choose the field against which you want to calculate the difference.
    In Defined Field Name enter a name, this will hold the value of the calculated difference.
    Thank You
    Shabaz

  • Source of data for custom tables in Oracle Apps

    Hi,
    Is there any way to find the source of data for the Custom tables. The source can be either any function, procedure, package or a trigger or any other process. The reason why I need to know about it is because we have some custom tables that are used by some custom Reports. Sometimes those reports comes out blank because the data in these custom tables is blank. So I need to know which processes are populating data in these custom tables. Any help would be really appreciated !!
    Thanks,
    KM

    Is there any way to find the source of data for the Custom tables. The source can be either any function, procedure, package or a trigger or any other process. The reason why I need to know about it is because we have some custom tables that are used by some custom Reports. Sometimes those reports comes out blank because the data in these custom tables is blank. So I need to know which processes are populating data in these custom tables. Any help would be really appreciated !!I do not think there is a direct way to find out the source of the data unless your custom processes are documented.
    You may enable AutditTrail (search the forum for links/docs) and see if this would help.
    Thanks,
    Hussein

  • Programmed Binding (Container IF) - how to access source container data

    Hi,
    I have created a programmed binding (container IF) using a function module based on SWA_BINDING_TEMPLATE.
    I want to access an element (Meldung) based on BUS2078 (QM notification) from my source container and then retrieve the material from this element.
    When the element is a simple element such as WfInitiator for example I cant retrieve the value using the get method on the source container but when it is a complex element such as a BOR object it does not work. How can I retrieve such data.
    Thanks for your help,
    Emmanuel

    I finally found the answer last week, here is how I have done :
        DATA : l_matnr TYPE matnr.
        DATA : l_string TYPE string.
        DATA: lcx_container TYPE REF TO cx_swf_cnt_container.
        DATA: b_meldung TYPE swf_bor_object.
        TRY.
            CALL METHOD source_container->get
              EXPORTING
                name  = 'Meldung'
              IMPORTING
                value = b_meldung.
            swf_get_property b_meldung 'Material' l_matnr.
          CATCH cx_swf_cnt_elem_not_found .
            l_string = 'cx_swf_cnt_elem_not_found'.
          CATCH cx_swf_cnt_elem_type_conflict .
            l_string = 'cx_swf_cnt_elem_type_conflict'.
          CATCH cx_swf_cnt_unit_type_conflict .
            l_string = 'cx_swf_cnt_unit_type_conflict'.
          CATCH cx_swf_cnt_container INTO lcx_container.
            l_string = lcx_container->t100_msg.
        ENDTRY.
    Emmanuel

  • No data selected in data source 2LIS_04_PECOMP

    Hi SAP BI Experts,
    We are facing issues while extracting data from infopackage u201CFULL /
    2LIS_04_PECOMP InfoCube 0PP_C08u201DData source u201C2LIS_04_PECOMPu201D.
    We have deleted the erroneous data from cube 0PP_C08 date range
    26.02.2009 to 2.03.2009. Now while extracting the data (Full load
    infopackage) from data source no record selected message displayed.
    In PSA no data selected. We have checked in RSA3 as well no data
    selected from 26.02.2009 to 03.03.2009.
    Please note that Delta Infopackage for same data source is working fine.
    Full Infopackage & RSA3 extracting data till Dec 2008 but 2009 no data available.
    Note : Action Taken
    DELTA is enabled on this data source, since the day we went LIVE. And because of the RECENT erroneous data deletion,
    We want to repair the data by running the FULL upload for the week.  But when we run the full repair upload info package,
    we get no result.   We tried running RSA3 for the same week, that also returned no records.
    RSA7 still shows the record ; since the DELTA is active. Hence the data is available in the queue.
    2.     Not sure of this but I tried to setup the statistical data using OLIFBW u2013 Reorg REP manufacturing ,  it gives the message,
    DataSource 2LIS_04_PECOMP contains data still to be transferred Message no. MCEX024.
    My log in NPRT does not show any records. I checked the setup table MC04P_0COMSETUP, there is no data.
    Kindly suggest the next step to get this data from 2LIS_04_PECOMP.
    Please help us on this.
    Thanks & Regards
    Abdul Karim

    Hi,
    and to fill the set up tables..
    Re: No data selected in data source 2LIS_04_PECOMP
    Re: No data selected in data source 2LIS_04_PECOMP
    rgds,

  • Xml as data source in Data Model

    Hi All,
    For testing, I am using sales.xml as a data source for creating the Data Model. There is a region field which I want to filter using the parameter. I have created the parameter, I am not able to use this parameter to filter the xml values. Is there a way to filter the xml output according to the parameter in BI publisher 11g while creating the data model
    Thanks,
    Virat
    Edited by: 929521 on Oct 16, 2012 2:13 PM

    can u please share the resolution

  • Error: 0xC020902A at Data Flow Task, XML Source [24515]: The "component "XML Source" (24515)" failed because truncation occurred, and the truncation row disposition on "output column "MsgLev1" (26196)" specifies failure on truncation. A truncation error o

    When I was Importing data from XML to SqlServer using SSIS , I am getting this error. The import is working if i use small file and not working if I use large XMl file. Can any one of you guys help me out with the issue?
    Error: 0xC020902A at Data Flow Task, XML Source [24515]: The "component "XML Source" (24515)" failed because truncation occurred, and the truncation row disposition on "output column "MsgLev1" (26196)" specifies failure on truncation. A truncation error occurred on the specified object of the specified component.
    Error: 0xC02092AF at Data Flow Task, XML Source [24515]: The component "XML Source" (24515) was unable to process the XML data. Pipeline component has returned HRESULT error code 0xC020902A from a method call.
    Error: 0xC0047038 at Data Flow Task: SSIS Error Code DTS_E_PRIMEOUTPUTFAILED.  The PrimeOutput method on component "XML Source" (24515) returned error code 0xC02092AF.  The component returned a failure code when the pipeline engine called PrimeOutput(). The meaning of the failure code is defined by the component, but the error is fatal and the pipeline stopped executing.  There may be error messages posted before this with more information about the failure.
    Error: 0xC0047021 at Data Flow Task: SSIS Error Code DTS_E_THREADFAILED.  Thread "SourceThread0" has exited with error code 0xC0047038.  There may be error messages posted before this with more information on why the thread has exited.
    Error: 0xC0047039 at Data Flow Task: SSIS Error Code DTS_E_THREADCANCELLED.  Thread "WorkThread0" received a shutdown signal and is terminating. The user requested a shutdown, or an error in another thread is causing the pipeline to shutdown.  There may be error messages posted before this with more information on why the thread was cancelled.
    Error: 0xC0047039 at Data Flow Task: SSIS Error Code DTS_E_THREADCANCELLED.  Thread "WorkThread1" received a shutdown signal and is terminating. The user requested a shutdown, or an error in another thread is causing the pipeline to shutdown.  There may be error messages posted before this with more information on why the thread was cancelled.
    Error: 0xC0047039 at Data Flow Task: SSIS Error Code DTS_E_THREADCANCELLED.  Thread "WorkThread3" received a shutdown signal and is terminating. The user requested a shutdown, or an error in another thread is causing the pipeline to shutdown.  There may be error messages posted before this with more information on why the thread was cancelled.
    Error: 0xC0047039 at Data Flow Task: SSIS Error Code DTS_E_THREADCANCELLED.  Thread "WorkThread4" received a shutdown signal and is terminating. The user requested a shutdown, or an error in another thread is causing the pipeline to shutdown.  There may be error messages posted before this with more information on why the thread was cancelled.
    Error: 0xC0047039 at Data Flow Task: SSIS Error Code DTS_E_THREADCANCELLED.  Thread "WorkThread2" received a shutdown signal and is terminating. The user requested a shutdown, or an error in another thread is causing the pipeline to shutdown.  There may be error messages posted before this with more information on why the thread was cancelled.
    Error: 0xC0047021 at Data Flow Task: SSIS Error Code DTS_E_THREADFAILED.  Thread "WorkThread0" has exited with error code 0xC0047039.  There may be error messages posted before this with more information on why the thread has exited.
    Error: 0xC0047021 at Data Flow Task: SSIS Error Code DTS_E_THREADFAILED.  Thread "WorkThread1" has exited with error code 0xC0047039.  There may be error messages posted before this with more information on why the thread has exited.
    Error: 0xC0047021 at Data Flow Task: SSIS Error Code DTS_E_THREADFAILED.  Thread "WorkThread2" has exited with error code 0xC0047039.  There may be error messages posted before this with more information on why the thread has exited.
    Error: 0xC0047021 at Data Flow Task: SSIS Error Code DTS_E_THREADFAILED.  Thread "WorkThread3" has exited with error code 0xC0047039.  There may be error messages posted before this with more information on why the thread has exited.
    Error: 0xC0047021 at Data Flow Task: SSIS Error Code DTS_E_THREADFAILED.  Thread "WorkThread4" has exited with error code 0xC0047039.  There may be error messages posted before this with more information on why the thread has exited.

    The reason is in the first line of the error.  It doesn't have anything to do with the size of your XML file - it has to do with the contents of the "MsgLev1" column.  You (or the XSD) has indicated that the values in this column not exceed a certain size - but one of the values in the file that's failing is larger than that.
    In order to fix the problem, you're going to need to increase the allocated space for that column.  In order to do that, you're going to need to find out what the required size of that data is - either from someone who ought to know, or by direct examination of the file.  If you want to know which entity has this overly large data element in it, you need to configure the error handling of the XML Source to "redirect truncation errors".  Then you can hook up the XML Source's error output to another destination where you can see which rows are problematic.
    Talk to me on

Maybe you are looking for