Using Webservice as data source in BI publisher data model

Hi All,
i want to expose data from share point through a webservice and use that webservice as a data source in Bi Publisher. But i am not able to use the webservice directly since it is a external web service.
For that i made below config changes. But it is causing weblogic server starting issues.
MW_HOME/user_projects/domains/DOMAIN_NAME/bin/
EXTRA_JAVA_PROPERTIES="-Dhttp.proxyHost=Hostname -Dhttp.proxyPort=80 -Dhttp.nonProxyHosts=localhost|hostname ${EXTRA_JAVA_PROPERTIES}"
export EXTRA_JAVA_PROPERTIES
EXTRA_JAVA_PROPERTIES="-Djavax.xml.soap.MessageFactory=oracle.j2ee.ws.saaj.soap.MessageFactoryImpl
-Djavax.xml.soap.SOAPFactory=oracle.j2ee.ws.saaj.SOAPFactoryImpl -Djavax.xml.soap.SOAPConnectionFactory=oracle.j2ee.ws.saaj.client.p2p.HttpSOAPConnectionFactory ${EXTRA_JAVA_PROPERTIES}"
export EXTRA_JAVA_PROPERTIES
Please let me know what i am doing is correct and suggest something.

i want to expose data from share point through a webservice and use that webservice as a data source in Bi Publisher. But i am not able to use the webservice directly since it is a external web service.
But i am not able to use the webservice directly since it is a external web service.
why?
look at "Creating a BI Publisher Report Based on External Web Services" ExternalWebServices
For that i made below config changes. But it is causing weblogic server starting issues.
why do you want to change some weblogic server settings?

Similar Messages

  • PowerPivot report data refresh error Data source as SharePoint list Data Feed

    Hi All,
    I am facing a problem on auto refresh, the report data source is SharePoint list Data Feed, I saved it my PC desktop when I set up auto fresh it fails then I moved the SharePoint list data feed "atomsvc" file to A SharePoint Data feed library then connected
    to report started a manual refresh it was success, after next schedules it failed.
    Error massage:
    Errors in the high-level relational engine. The following exception occurred while the managed IDbConnection interface was being used: 
    The network path was not found. ;The network path was not found. The network path was not found. . 
    A connection could not be made to the data source with the DataSourceID of '82f49f39-45aa-4d61-a15c-90ebbc5656d', 
    Name of 'DataFeed testingreport'. An error occurred while processing the 'Testing Report' table. The operation has been cancelled.  
    Thank you very much !
    erkindunya

    If you are using Claims Auth then this is a known limitation.  Auto refresh only works under Classic Auth.
    I trust that answers your question...
    Thanks
    C
    http://www.cjvandyk.com/blog |
    LinkedIn | Facebook |
    Twitter | Quix Utilities for SharePoint |
    Codeplex

  • Using an SAS2008 Cube as a data source for BI Publisher errors

    Hi All,
    I am trying to use a SAS 2008 cube as a data source for a report in BI Publisher. When I defined the data source under the admin section I gave it the following details:
    Data Source=ntsasqa01;Provider=msolap.3;Initial Catalog=WWPR_SMG
    With a username / password that has access to the cube.
    When I try and test the connection the OC4J instance stops running and the test does not finish or error. If I use the data source and then setup a report and then try and view data I get an error like the one below:
    /app/oracle/OracleBI/xmlp/XMLP/Admin/DataSource/msmdacc.dll (     0509-022 Cannot load module /app/oracle/OracleBI/xmlp/XMLP/Admin/DataSource/msmdacc.dll.
         0509-103 The module has an invalid magic number.)
    I am not sure if the error is related to the initial issue of not being able to complete the test.
    I need to find a solution though as I have to make use of MDX to access the cubes.
    Please could someone give me some advice.
    Thanks,
    Kim.

    Hi Kim
    1. SAS cubes are not currently supported - http://download.oracle.com/docs/cd/E10415_01/doc/bi.1013/e10417.pdf. Unless SAS cubes can use the MS drivers?
    2. If you try and pursue it you might check the permissions on the ddll in the error to ensure you have rights over it.
    Either way you are not supported with a SAS cube so you are kinda on your own. Sorry I dont use SAS so Im not going to be much help other than stating the obvious :0)
    Regards
    Tim

  • Multiple Data Sources for XML Publisher in APPS?

    Is it possible to draw data from two different reports (i.e. two different data sources) into a single template, using XML Publisher in APPS?

    Gareth is correct if it's only ONE report. It's my understanding that it's two reports that you want to merge into one report. If this is the case then the options I gave you are still correct.
    Option 3: You could add a third option which is to keep the reports the same, set the output in the concurrent program(s) to xml. You will still have to merge the xml files and apply the format template which would require a javaConcurrentProgram.
    Option 4. You could do option three but with a twist. Create two format templates, apply them to each xml output, then use the pdfMerger utility to create a single document. This may or may not work based on your requirements.
    Ike Wiggins
    http://bipublisher.blogspot.com

  • Issue with SQL Query with Presentation Variable as Data Source in BI Publisher

    Hello All
    I have an issue with creating BIP report based on OBIEE reports which is done using direct SQL. There is this one report in OBIEE dashboard, which is written using direct SQL. To create the pixel perfect version of this report, I am creating BIP data model using SQL Query as data source. The physical query that is used to create OBIEE report has several presentation variables in its where clause.
    select TILE4,max(APPTS), 'Top Count' from
    SELECT c5 as division,nvl(DECODE (C2,0,0,(c1/c2)*100),0) AS APPTS,NTILE (4) OVER ( ORDER BY nvl(DECODE (C2,0,0,(c1/c2)*100),0))  AS TILE4,
    c4 as dept,c6 as month FROM 
    select sum(case  when T6736.TYPE = 'ATM' then T7608.COUNT end ) as c1,
         sum(case  when T6736.TYPE in ('Call Center', 'LSM') then T7608.CONFIRMED_COUNT end ) as c2,
         T802.NAME_LEVEL_6 as c3,
         T802.NAME_LEVEL_1 as c4,
         T6172.CALENDARMONTHNAMEANDYEAR as c5,
         T6172.CALENDARMONTHNUMBERINYEAR as c6,
         T802.DEPT_CODE as c7
    from
         DW_date_DIM T6736 /* z_dim_date */ ,
         DW_MONTH_DIM T6172 /* z_dim_month */ ,
         DW_GEOS_DIM T802 /* z_dim_dept_geo_hierarchy */ ,
         DW_Count_MONTH_AGG T7608 /* z_fact_Count_month_agg */
    where  ( T802.DEpt_CODE = T7608.DEPT_CODE and T802.NAME_LEVEL_1 =  '@{PV_D}{RSD}' 
    and T802.CALENDARMONTHNAMEANDYEAR = 'July 2013'
    and T6172.MONTH_KEY = T7608.MONTH_KEY and T6736.DATE_KEY = T7608.DATE_KEY
    and (T6172.CALENDARMONTHNUMBERINYEAR between substr('@{Month_Start}',0,6)  and substr('@{Month_END}',8,13))
    and (T6736.TYPE in ('Call Center', 'LSM')) )
    group by T802.DEPT_CODE, T802.NAME_LEVEL_6, T802.NAME_LEVEL_1, T6172.CALENDARMONTHNAMEANDYEAR, T6172.CALENDARMONTHNUMBERINYEAR
    order by c4, c3, c6, c7, c5
    ))where tile4=3 group by tile4
    When I try to view data after creating the data set, I get the following error:
    Failed to load XML
    XML Parsing Error: mismatched tag. Expected: . Location: http://172.20.17.142:9704/xmlpserver/servlet/xdo Line Number 2, Column 580:
    Now when I remove those Presention variables (@{PV1}, @{PV2}) in the query with some hard coded values, it is working fine.
    So I know it is the PV that's causing this error.
    How can I work around it?
    There is no way to create equivalent report without using the direct sql..
    Thanks in advance

    I have found a solution to this problem after some more investigation. PowerQuery does not support to use SQL statement as source for Teradata (possibly same for other sources as well). This is "by design" according to Microsoft. Hence the problem
    is not because different PowerQuery versions as mentioned above. When designing the query in PowerQuery in Excel make sure to use the interface/navigation to create the query/select tables and NOT a SQL statement. The SQL statement as source works fine on
    a client machine but not when scheduling it in Power BI in the cloud. I would like to see that the functionality within PowerQuery and Excel should be the same as in Power BI in the cloud. And at least when there is a difference it would be nice with documentation
    or more descriptive errors.
    //Jonas 

  • Why do SQL sample programs work when there is no entry for the data sources in the ODBC Data Source Administrator that are used by the sample program?

    Hi,
    I am trying to understand how to access an SQL database from CVI.  
    I don't understand how the sample programs shipped with the toolkit can access the database they use if it is not listed in the ODBC Data Source Administrator.
    In the code, I see the line:
    hdbc = DBConnect ("DSN=CVI SQL Samples");
    However, there is not an entry that I can find in the ODBC Administrator for "CVI SQL Samples."
    The SQL help suggests there should be a User DSN named, "CVI32_Samples", but I don't see that entry in the ODBC Administrator on my win-7 system.
    I believe the sample program is accessing a database file named, "C:\Users\Public\Documents\National Instruments\CVI\Samples\sql\samples.mdb".
    Exactly how is the call to DBConnect() translated to the file, "samples.MDB", if there is no entry in the ODBC Administrator?
    Regards,
    Mark 

    I figured it out.  There is a 32 bit ODBC administrator and a 64 bit ODBC administrator.  The ODBC administrator I run through the control panel is the 64 bit ODBC administrator. To access the 32 bit administrator I need to run, "C:\Windows\SysWOW64\odbcad32.exe".
    I am running the samples in 32 bit mode, so I need to use the 32 bit ODBC administrator.
    Best,
    Mark

  • Data Source to upload plan data for CO_OM_CCA_: CC: Costs and Allocations

    Hi Guru's,
    We have Data source that which upload Actual data for CCA (0CO_OM_CCA_1 - Cost Centers:cost and allocations).  Every time it is full upload before loading the data deletion of previous load request will be done.
    One more information required that i have checked in the BW cube for plan data it is available in the info cube for one cost center and rest of the costcenter i don't have the data now the users are requesting to upload the plan data for rest of cost centers.
    Now as per the business requirement the users are requesting to upload the plan data in to the same i am not sure whether it will pull the plan data with actuals in to BW cube. Is there any alternate with the same data source or else is there any specific data source for uploading Costs and Allocations plan data
    Please suggest me how can i go a head with this requirement.
    Thanks in Advance.
    Ganni

    You can use the same datasource to load plan/budget data to the cube for all the costcenters. And regarding the actuals, you can use 0CO_OM_CCA_9 datasource to load to actuals cube and create a multi on top of both plan and actuals cube for reporting purposes.

  • External Data source not appearing in Data source type

    Hello,
    I use SQL Server 2012 BI Edition and use Data tools to create reports. I have a 3rd party Data source. While creating a new shared data source, I cannot get it to appear under the Data Source Type. Only the current built in types are available. For test
    purposes, I tried editing the RSReportServer.config and added a duplicate entry and it was shown in the dropdown. What should I do? 
    When this 3rd party data source is installed, under Tools --> Connect to Database, I am able to get that datatype and connect to my data source. I tried uninstalling and this option goes away. 
    Regards
    Vivek.

    Hi Vivek,
    Based on my research, the error always occurs when you load custom extensions in SQL Server 2012 Data Tools or use Report Builder 2.0 to run an SSRS 2008 or SSRS 2008 R2 report that uses a custom data processing extension in server mode. According to your
    description, it seems that you are not in those two scenarios. So I think this issue can be caused by that the Custom Data Process Extension is not deployed correctly in SQL Server Reporting Services.
    The following document about Implementing a Data Processing Extension in Reporting Services is for your reference:
    http://www.codeproject.com/Articles/22946/Implementing-a-Data-Processing-Extension
    Besides, please note that Report Builder is currently not extensible, and can't load custom data extensions or custom query designers. We can refer to the method posted by Robert to work around this issue:
    https://go4answers.webhost4life.com/Example/custom-data-sources-report-builder-30-15682.aspx
    Hope this helps.
    Thanks,
    Katherine Xiong
    Katherine Xiong
    TechNet Community Support

  • STANDARD DATA SOURCE 2LIS_11_VAHDR NOT GETING DATA

    RESPECTED GURUS
    I AM A NEW BI USER. I AM TRYING TO EXTRACT DATA FROM 2LIS_11_VAHDR WHICH IS A STANDARD DATASOURCE THROUGH RSA3, BUT I AM GETTING A MESSAGE "0 RECORD FOUND" BUT THERE ARE PLENTY OF DATA. I WAS TRIED WITH THE OTHER STANDARD DATA SOURCES AND GETTING THE SAME MESSAGE" ZERO RECORD FOUND".PLS GUIDE ME HOW I WOULD I GET DATA.
    REGARDS
    ABHAY MAHODAYA

    Hi Abhay,
    The data source which you are trying to use is comes under logistics. To take the logistics data from R/3 to BW set up table needs to be filled up. To fill the set up table follow the below mentioned procedures.
    First Delete the set up table to avoid data duplication:
    Go to T.Code: SBIW -> Settings for Application Specific data sources (PI) -> Logistics -> Managing Extract structures -> Initialization -> Delete the contents of the setup table
    Fill the set up table:
    Go to T.Code: SBIW -> Settings for Application Specific data sources (PI) -> Logistics -> Managing Extract structures -> Initialization -> Filling in the setup table -> Application specific set up of statistical data
    Choose the relevant node to perform the set up based on your requirement.
    Remember while filling the set up table no entry sould get posted in the relevant tables. So while you do this in production system, you may have to request for user locking for the relevant T.Codes.
    All the best.
    Regards,
    Sarath.

  • Design Studio 1.3 : one large generic Data Source or multiple smaller Data Sources

    Dear all,
    In DS 1.3, is it still a best practice to have one large generic Data Source ? Or is having multiple smaller Data Sources a better solution ?
    Minimizing the number of Data Sources remains a golden rule, but what is the best solution :
    One large generic Data Source : and using setDataSource
    or
    Multiple smaller Data Sources : and using Load in Script and Background Processing
    Many thanks for sharing your ideas,
    Hans

    It depends on your application and how much the data is pulling in and how you want to present that to the user/consumer of the application.
    At TechEd Las Vegas last year, SAP showed a 9 dashboards (3 per row) with background processing for each row.

  • Generic Data Source with Function Module data mismatch in BI

    Hi All,
    I'm using Generic Data Source with Function Module, When I execute the Function Module (Which I have Created), I'm getting 16000 records and when run extractor(in RSA3) im getting different no.of records(infact they are more no.).
    when I run the InfoPackage in BI im Getting more no. of records than what i got executing the function module..
    and single record is divided into 2 records in BI side(not all the records), how can it be possible???
    is there anything Im missing to explain you my issue???
    if understood please help me out.
    Thanks n Regards,
    ravi.

    the datasource frame work starts the function module several times.
    1. the initialization
    2. the serval times, until you "raise no_more_data".
    check you coding: have you refreshed necessary internal tables.
    Sven

  • Generic Data Source with Function Module data mismatch

    Hi All,
    I'm using Generic Data Source with Function Module, When I execute the Function Module (Which I have Created), I'm getting 16000 records and when run extractor(in RSA3) im getting different no.of records(infact they are more no.).
    when I run the InfoPackage  in BI im Getting more no. of records than what i got executing the function module..
    and single record is divided into 2 records in BI side(not all the records), how can it be possible???
    is there anything Im missing to explain you my issue???
    if understood please help me out.
    Thanks n Regards,
    ravi.

    HI rkiranbi,
    1. FIrst you excute function module according to your paramers, you will get some records. then goto tcode RSA3 --> excute
    Provide your Data source name and under setting we have options like Data records/calls, Display extractor calls and selections --> fields .
    in that options you have to increase the values. and then you have to pass paramers in RSA3 according to your function module
    selections in SE37. Now you will get equal values in both functin module selection and RSA3 Selection. if it fail means  you need to
    check coding logic in function module. 
    2. if your  are getting wrong values in BI System then check with
                  1. compare with PSA data and data target data (here you need to check with characterstic as well as keyfigures)
                  if you find any mistake you need change the coding in function module according to client requirement.
                  2. compare data with RSA3 and bi report data or data target data.
                                 check it properly above steps, you will get solution.
    thanks and regards,
    malli

  • What is the data source name and the data target name for the table COSP

    Hi,
    Actaully i am new to fico/bw,and i have to create a report based on the actual and budget value and difference in the variance.
    in R/3 side the table they are using is COSP,so please let me know what is the data source name and the cube or ods name in BW side.wht are the fields in which i can get the BUDGET value in the report.
    here we are using all the business content extractors and BI content cubes and odss.
    please reply immediately as its very jurgent isssue.
    Thanks,
    ashok

    answered

  • Dashboard when published to BW, date functions always show published date

    Xcelsius dashboard when published to BW at a particular date, the date functions(like Today(), now()) used at spreadsheet generates published date irrespective of time or date the dashboard is previewed....Could anyone help me out what actually .XLF gets stored in BW....does these functions stores a static value when published?

    Hi Rajesh,
    To confirm, you're saying that you're saving a swf (falsh file) to BW and when it's opened by a user in a browser, it doesn't show the most recent date?
    In the current version, the Now() function will be calculated at the time the dashboard is opened. Each time you open the swf file you will see the current day/time.
    You mention the xlf file. The date stored in this has nothing to do with what is viewed in the swf file (if you're in BI4.0 then the xlf and swf are stored in the same object on the server).

  • Combining heterogeneous data sources in a single data target

    Dear Experts,
    I have a requirement as below for combining multiple data sources (different key fields) into a single target
    Target Key Fields : sales doc, item, schdl line, delivery no and delivery item.
    1 source key fields : sales doc and item  (VAITM)
    2 source key fields :sales doc, item, schdl line, delivery no and delivery item (V_SSL)
    3. source key fields : delivery number and delivery item. (VCITM)
    Is there an innovative way of combining all these records in a single record in data target rather then creating multiple entries (due to different source keys) rather than using a conventional look up routines.
    We are on BI 7.0.
    Thanks,
    Rita
    Edited by: Rita Tripathi on Jan 19, 2012 7:05 PM

    Hi,
    Target Key Fields : sales doc, item, schdl line, delivery no and delivery item.
    1 source key fields : sales doc and item (VAITM)
    2 source key fields :sales doc, item, schdl line, delivery no and delivery item (V_SSL)
    3. source key fields : delivery number and delivery item. (VCITM)
    You can actually create a View on (1) and (2) tables in SE11 and then create a generic datasource in RSO2.
    You create one more generic DS from (3) table.
    Combine these two DS in an Infosource and update to DSO.
    Regards,
    Suman

Maybe you are looking for

  • How to add function group entry in Transport request...

    Hi I had a function group in which I had 4 function modules. I deleted one of the FM and hence in my Transport request, its entry was included. Along with that the function group was also included. I deleted that function group entry by mistake. Now

  • OBIEE 11G Configuration??

    Hi Experts, I tried installing OBIEE 11g. During configuration all of a sudden my computer turned off, now is there any way that we can resume the installation or at least is there a way to do only configuration?? Thanks, DK

  • Syncing photos from network drive

    all my photos are stored on a network drive and itunes successfully syncs all photos from this drive to my iphone. however, the next time i turn on my computer and re-sync, i see from the photos tab that the location of my photos has been changed to

  • Why canu0092t I enable capture of statistics performance data?

    Hi, Following some documents suggested here, I tried to enable capture of statistics performance data. This is what I did: rsa1, Tools (from menu) Settings for BI statistics Then I selected the Infoprovider tab; Under the “statistics” column, it look

  • Why is /etc/services full of email addresses?

    No other Unix system (I've ever used) stuffs email addresses between each port group (in the comments). What's the deal?