Alternative Data Sources

Has anyone found or figured out a way to connect spry to
alternative datasources other than XML?
I'm just want something simple like either converting a sql
query into an XML format so spry can recognize it or something
similar.
I'm using asp pages with javascript.
Any thoughts or suggestions would be greatly appreciated.
Michael Lee

Hi Michael,
We are working on JSON and HTML data sets, but for now, it's
XML.
We have a sample file for turning ASP queries into XML:
http://labs.adobe.com/technologies/spry/samples/utils/query2xml.html
Hope this helps,
Don

Similar Messages

  • Use View Object with Alternative Data Sources in View Link

    In my case, I created a view object based on alternative data source (not with SQL query). I have setQuery(null) in the override create() function for this View Object. When I use this View Object at the destination end of the View Link, the setQuery(null) will raise Jbo-26016 exception (You cannot set customer query (calling setQuery()) on a view object if it is the detail view object in a master detail view link.) which does not allow me to do that. I looked Avrom's Framework for Database API-Based ADF BC, in which he override the DBTransaction and CallableStatement in order to intercept the default query clause set by the View Link. I'd like to see if there is any alternative ways to achieve my objective?

    Just following up what I found so far. I'm trying to use customized View Object based on alternative data source at the destination end of View Link with ADF 10.1.3.
    1. For the View Object at the destination end of View Link, I can not call setQuery() which will raise Jbo-26016 exception.
    2. I can omit the setQuery(null) call in my customized ViewObjectImpl, and as long as there is no SQL statement in the View Definition XML, it won't cause problem when calling the executeQueryForCollection() of the super class.
    3. The override executeQueryForCollection() must call the super.executeQueryForCollection() in order to get appropriate result (I tried to skip the call and it does not give back any results)
    4. The funny thing is, when I make getQuery() call in my customized ViewObjectImpl (I'm just trying to double check if the Query object is clean or not), the query object is set to "SELECT FROM" following the getQuery() call, and which causes the exceptions when I call super.exeuteQueryForCollection(). I found out later from the Javadoc API which is implemented in that way.
    So I guess my case is resolved by not making specific call to setQuery(null).

  • Alternative data source for EO and VO

    It is usually the requirement to customize the Entity Objects and View Objects so that they can work with data sources different than database schema objects (tables, views,...), some examples are Web Service or Oracle Coherence cache.
    I have developed mini projects using both data sources mentioned, by overriding relevant methods in the implementation EO and VO classes. There is one limitation I realize when developing these projects. Although both the EO and VO rely directly and completely on the alternative data sources (Web Service and Coherence), still there must be a schema object existing in the local database to back the EO and VO. This is because I still have to call super methods from ViewObjImpl and EntityImpl, and these methods are still doing some operations with the schema objects, and without these objects, ORA-00942: table or view does not exist will be thrown.
    For example, after overriding executeQueryForCollection in ViewObjImpl to take data from web service, it is still necessary to call super.executeQueryForCollection so that data will actually appear on the web page, and without the schema objects, this method will throw ORA-00942.
    My question is, is there any way to completely eliminate the need of schema objects in EO and VO? I know the answer (if exist) will be complicated but I'm willing to explore.
    Thanks.

    Thanks Timo,
    I figured out how to achieve this.
    Basically when creating the EO, I did not specify the schema object in the first step, and for each attribute that I created in the wizard, I also unchecked the checkbox "persistent". I could then create VO based on this EO and override all the necessary methods in the EO and VO so they can do CRUD operations via Web Service.
    In the case of only creating an VO (read-only, for retrieving/reading data only), I could also mark each attribute as "transient" by checking a checkbox in the creation wizard.
    Edited by: nhd on May 7, 2013 8:01 PM

  • Excel-2007 cannot connect to Oracle ODBC data source, Control Panel can.

    <p>
    I cannot make an ODBC connection from Exce-2007 to Oracle work. I am an expert Excel and VBA user (since 1994) and I have frequently used Excel to access ODBC databases, including Oracle (I have done this with Excel-2003 both with worksheet queries and have written VBA ADO-connection routines). And even though in Excel-2007 a worksheet ODBC query is supposed to be easier to create than in previous version of Excel, my connection fails. Any suggestions and all help are welcome and much appreciated.
    DETAILS
    </p>
    <p>
    1) <strong>What is my system?</strong> I am using Excel-2007 on Windows Vista x64 and Oracle server v.11g on my computer (all this is on my computer, no network issues).
    2)<strong> Why use Excel with Oracle at all?</strong> I use Excel-2007 to access Oracle rather than Access-2007 (or any other application like TOAD, etc.) because I do engineering calculations with the data stored in Oracle. These calculations are easier done in Excel (I suppose that one alternative to this could be to use some sql or Access to get the data from the database, then store it as plain vanilla CSV file, then open this file in Excel, then do the math (the math involves complex optimisation algorithms), then save the results as CSV, then use some sql or Access to put the data back into the database. Howwever this does not strike me as a quick or neat solution. And after all Excel has been designed to access ODBC databases, so why not use it?)
    3) <strong>What do I do in Excel-2007 that won't work?</strong> I create an ODBC link to Oracle that does not work. In Excel-2007 this is straightforward:
    </p>
    <ul>
         <li>define an ODBC connection (Data tab --&gt; From other sources --&gt; From data connection wizard);</li>
         <li>define a query on the worksheet -- that's it, this is all!</li>
    </ul>
    <p>
    I start with creating an ODBC connection:
    a) I choose an ODBC data source type: <strong><em>ODBC DSN</em></strong>
    b) Excel-2007 displays the list of the available ODBC data sources. I see in it <strong><em>my Oracle database name</em></strong> and I select it.
    c) Excel-2007 displays the Data Link Properties:
    - the "Provider" has a list of the OLE DB drivers with preselected "<strong><em>Microsoft OLE DB Provider for ODBC Drivers</em></strong>". I keep this default selection.
    - the "Connection" tab has a connection string "<strong><em>DSN=&lt;my database name&gt;</em></strong>" which I keep, it also has fields for the <strong><em>user name</em></strong> and the <strong><em>password</em></strong>, which I fill with the correct credentials.
    - Finally there is a button "Test Connection", which when I click produces the following error message:
    <strong><font color="#ff0000">"Test connection failed because of an error in initializing provider.</font></strong><strong><br />
    </strong><strong><font color="#ff0000">Unespecified error"</font></strong><strong>
    </strong>
    4) <strong>Additional food for thought:</strong>
    a) In the above walk-through the only data, which I type, are the user name and password, everything else is selected from lists offered by Excel-2007, hence any possibility of typos being the cause of the problem can safely be discarded.
    b) I can test the ODBC driver in the Control Panel and it shows that it can connect to the Oracle database:
    - in <strong><em>Control Panel --&gt; Admin Tools --&gt; Data Sources (ODBC)</em></strong> on the "User DSN" tab I can see the list of the available ODBC data sources (same list as in Excel-2007, point 3b above) with the name of my database in it;
    - selecting the name of my database from the list of the sources and clicking "Configure" button opens a tab with <strong><em>Data Source Name</em></strong> (same as in Excel-2007), TNS Service Name and User ID. I enter <strong><em>&lt;user name&gt;/&lt;password&gt;</em></strong> and click "Test Connection" button. A message "<strong><em>Connection successful</em></strong>" appears (just for a test I enter <u>incorrect user credentials</u> and "<strong><em>Unable to connect</em></strong>" message appears)
    BOTTOM LINE
    </p>
    <p>
    The procedure for using an ODBC connection from Excel is very simple, in the past I have created and used such connections numerous times with Excel-2003 and earlier on Win-XP and earlier. But now on Excel-2007 and Vista-x64 I cannot make it work.
    Also, testing an ODBC connection driver is really easy and simple to be done in the Control Panel. There testing the same ODBC connection, which fails in Excel-2007, results in success.
    I am frustrated by the simplicity of the problem and yet the persistant error. I have lost now two full days in failed attempts to make the simple procedure work and in searching the internet for answers.
    All help is highly appreciated <img class="emoticon" src="images/emoticons/happy.gif" border="0" alt="" width="16" height="16" />,
    Plamen
    </p>

    Did you find the solution to your problem?
    If not, I think I may know.
    Excel 2007 is a 32-bit application.
    When installing 32-bit applications in a 64-bit environment, the "default" location is:
    C:\Program files (x86)\...
    Excel then launches from this location.
    However, when it connects to Oracle and passes the name of the calling program, Oracle attempts to "interpret" the value of (x86) as if
    the value within the parenthesis are being passed as a reference. Of course, iOracle doesn't find anything, so the result is (),
    and then it cannot return the connection info back to the calling application.
    I corrected it by installing Excel in C:\Program Files\, and once launched from that location, it works the same as on the 32-bit machines.
    However, at my location, they are FORCING Excel to be installed in the(x86) location.
    What I'm trying to discover now is:
    Is it possible to flag Oracle to NOT process the embedded variables?
    Or, is it possible to assign a variable in Oracle such that x86 = "(x86)", so that the end result is viable?
    Have you had any luck with your installation?
    thanks,
    Paul

  • Reg-Creating Data source in BI

    Hi All,
    I need to crate a data source in BI By using the following tables TBTCP, TBTCO, TRDIRT, TSTC..First two tables having common relationship, where as there is no relationship between TRDIRT and TSTC tables....For this i am planning to do by creating a Functional Module.
    Using Functional Module any draw back's are there?
    Let me suggest is there any alternative like BADI, BAPI........
    With Regards
    Aswin

    Since there is no common field in the rest of the two tables, you wont be able to make view on them. FM is the option. There is no drawback, the only thing is that you need to write the code.
    Thanks...
    Shambhu

  • Error while testing Generic Data Source extraction

    I've created generic data source for Texts and Attributes in R/3 based on VIEW (Z table)
    I get  error "Error 6 in function module RSS_PROGRAM_GENERATE" while trying to test the data sources through RSO2
    I guess there can be a lot of reasons... Please name me anybody who knows them...
    thanx in advance

    Hi,
    Please Check the OSS note 328948 and it provides solution to same error what you have. Additionally you can have a look into this OSS note 705212 too.
    Syntax error in the generated extraction program Symptom
    You notice the error by one of the following symptoms:
    The loading of transaction data from an R/3 system terminates with a syntax error in the generated extraction program. The monitor displays the error messages:
    R3027 "Error & during the generation of the data transfer program"
    RSM340 "Error in the source system"
    The extraction within the extractor checker terminates with error message RJ028 "Error 6 in function module RSS_PROGRAM_GENERATE".The activation of transfer rules ends in BW with error message RG102 "Syntax error in GP$ERR$, row... (-> long text)" from the source system. Usually, the diagnosis in the long text of the error message is: "...could not be interpreted. Possible error causes: Incorrect notation or... "
    Other terms
    OLTP, extractor, data extraction, DataSource, Service API, SAPI,
    R3 027, R3 27, RSM 340, RJ 028, RJ 28
    Reason and Prerequisites
    The error only occurs in the source system, if this contains Basis Release 3.1I and Service API (SAPI) 3.0C Support Package 6. SAPI 3.0C Support Package 6 is contained, for example, in PI 2003.1 Support Package 7 (see attached composite SAP note 673002).
    Solution
    To correct the problem you need service API 3.0C Support Package 7 in the affected source system. The attached composite SAP note 704971 explains in which software components service API 3.0C is contained, and what the corresponding Support Packages of these components are.
    Alternatively, you can also copy the advance correction from the appendix.
    Hope it helps.
    Regards

  • Issue with Crystal Report based on 2 different data sources

    Hi there,
    I am having a frustrating problem with a report I've designed and I'm hoping someone might be able to assist please.
    The report has 3 different prompts, each of which is based on a dynamic list of values retrieved via views within a SQL server db.
    We are wanting to introduce the use of Universes as much as possible, so the data returned is based off a BO Universe data source query.
    I have uploaded the report into BO and have provided the necessary database logon information for the report (in the "Process" > "Database" settings in the CMC) for both the direct db datasource connection (for the views) and the BO Universe query connection.
    When the report is run however, the report still prompts for the database user name & password credentials. I have triple checked my db connection settings, and also have "Use same database logon as when report is run" set to true. I also tested a cut-down version of the report without the Universe connection with the same db logon credentials I provided and there was no credentials prompt when it was run, proving those values are accepted.
    Does anyone know why this is happening & if there is a way around it? Alternatively, is there some way that a report prompt may be based on a dynamic list of values retrieved via a Universe connection? This way I'd be able to remove the db connection for the views and have the report solely based on the Universe.
    Another issue that occurs is out of the 3 prompts, a user can select a User Name OR Number, and also must select a Period. However if the User Name or Number is left blank the message "The value is not valid" is shown. So I tried a cut-down version of the report with only the BO Universe as a data source (static prompts) and this didn't occur, i.e. I was allowed to leave either the User Name or Number empty.
    I hope this all of makes sense - let me know if not. If anyone is able to help out with any of this it would be very much appreciated.
    Cheers,
    Marco

    Please re-post if this is still an issue to the Business Objects Forum or purchase a case and have a dedicated support engineer work with you directly

  • Data source creation

    Hi all,
    In general info tab in data source creation(BI 7.0) there is a option of delievery of duplicate records and it hae 3 option and they are undefined, none, allowed.so my question is how undefined is diff. from the other two?

    Hi Neha,
    This indicator gives information on how the DataSource behaves within a request with regard to duplicate records:
    ' ' The status of the DataSource is unknown.
    '0' The DataSource does not deliver any duplicate records within a request, with reference to its key.
    '1' The DataSource can deliver duplicate records within a request, with reference to its key. However, no duplicate records are delivered in a data package.
    This indicator is particularly important for delta-capable attribute tables and text tables.
    For the settings '0' and '1' you also have to define a key for the DataSource. This can be either in the DDIC using the maintenance for the corresponding field property of the extract structure fields, or (alternatively or additionally) in the metadata of the Datasource. A field in the DataSource also has the additional attribute ' DataSource Key Field', which transfersor corrects the DDIC property where necessary.
    Use
    For a key of time-independent master data or time-independent texts, DataSources can transfer several data records to BI in a request. If data records are transferred to the BI system more than once within a request, this can be the result of the application and you should not see it as an error. In a master data update, the BI system provides functions to handle duplicate data records. When data is loaded into another InfoProvider, the relevant handling must be coded manually, preferably in the end routine for the transformation.
    Hope its clear now...
    Regards
    Gattu

  • Populating a subform from a data source.

    I understand that the help file in LiveCycle Designer describes the procedure for connecting fields in a form to columns in a data base. I have tried for several days now to populate simple customer information (name, address, state, zip, phone, fax) from a data base.
    I have exported the data into an acess database, created the dsn, created the data connection, an secured the database. I am still getting an error message that says "Connection for Source Data Connection failed because the environment is not trusted." I am trying everything I can think of to create the proper combination of user-level security for the database and proper use of the "connection string", but now I have read in the forums that even upon success of this effort, my fellow employees will not be able to use the form properly unless we purchase "Reader Extension Server"
    Is there any way to populate customer data in a form, from a data source, that is compatible with Reader, without buying "extension server"
    Thank you in advance for your support.
    Regards,
    Greg Mitchell
    Chadwick Inc.

    Here is a quote from the LiveCycle Designer 7.0 Help file
    OLEDB is the only database connection that you can connect to from Designer. Through the use of the Microsoft OLEDB Provider for ODBC Drivers, you can also access databases identified by the ODBC data source name (DSN). You can use OLE drivers other than the Microsoft OLEDB Provider for ODBC, however the form must be certified in Acrobat before it will function at run time.
    In Acrobat, security concerns dictate that you cannot specify an ODBC connection string by using the Driver=; syntax. Therefore, the client computer needs a DSN configured for ODBC connections.
    If you are using external data sources in a form design that you want users to fill in Adobe Reader, you must set
    usage rights for the form by using Reader Extensions Server.
    Unfortunately it looks like a direct DB connection requires Reader Extensions. Fortunately you can get Reader Extensions pretty cheap from
    FormRouter.
    As an alternative you can submit the data to a server script (like ASP) and have it write into the Database. There are plenty of sources for writing ASP for PDF files. Just do a search on this forum, or on the web in general.
    Related Forum message:
    One site that seems pretty decent is:
    http://www.nk-inc.com/pdfsupport/pdf_programming.htm
    Thom Parker
    WindJack Solutions
    www.windjack.com

  • ODBC Data Source Won't Recover from Connection Errors

    I have several ODBC data sources to a mainframe and want to
    use the "Maintain Connections" setting as it dramatically improves
    performance under heavy load. Problem is that we frequently have
    path problems between the CF server and the mainframe 180 miles
    away. With the Maintain Connections checked, the data source
    doesn't consistently start working again after the path problem
    clears up. The path problem usually only lasts a minute or two, but
    SOME queries through the affected data sources continue to raise
    errors until the ODBC Service is cycled. The SOME aspect is really
    weird because not every hit through that data source reports an
    error after the path comes back up. I've had side-by-side users hit
    the same page using that DS and one consistently get's an error and
    the other works fine. Restarting the ODBC service clears the
    problem so it works fine for everybody. Of course that breaks all
    the connections that were still working so I just affected a bunch
    more users. Silver lining there is that at least nobody can claim
    preferential treatment right. It's also kind of bizzare that I can
    use the CF Administrator to verify all data sources and they
    usually all report that they are working fine even while some
    connections are still reporting problems.
    The specific ODBC drivers we are using are Neon Shadow
    Drivers (Multi Threaded). I'm thinking about using the Admin API to
    hit the Disable Connections and then re-enable it.
    Any constructive, tactful input for addressing this problem
    would be appreciated.

    spikehenning wrote:
    > I have several ODBC data sources to a mainframe and want
    to use the "Maintain
    > Connections" setting as it dramatically improves
    performance under heavy load.
    > Problem is that we frequently have path problems between
    the CF server and the
    > mainframe 180 miles away. With the Maintain Connections
    checked, the data
    > source doesn't consistently start working again after
    the path problem clears
    > up. The path problem usually only lasts a minute or two,
    but SOME queries
    > through the affected data sources continue to raise
    errors until the ODBC
    > Service is cycled. The SOME aspect is really weird
    because not every hit
    > through that data source reports an error after the path
    comes back up. I've
    > had side-by-side users hit the same page using that DS
    and one consistently
    > get's an error and the other works fine. Restarting the
    ODBC service clears the
    > problem so it works fine for everybody. Of course that
    breaks all the
    > connections that were still working so I just affected a
    bunch more users.
    > Silver lining there is that at least nobody can claim
    preferential treatment
    > right. It's also kind of bizzare that I can use the CF
    Administrator to verify
    > all data sources and they usually all report that they
    are working fine even
    > while some connections are still reporting problems.
    >
    > The specific ODBC drivers we are using are Neon Shadow
    Drivers (Multi
    > Threaded). I'm thinking about using the Admin API to hit
    the Disable
    > Connections and then re-enable it.
    Do you have a JDBC Type 4 alternative for these drivers? I
    got rid of
    all my datasource problems long ago when I got rid of ODBC
    altogether.
    Jochem
    Jochem van Dieten
    Adobe Community Expert for ColdFusion

  • Can you include an external data source in query condition?

    Hi all,
    I have a user that wants to run a report off a subset keys in an external data source. Is this possible in Disco Plus?
    For example, he has an Excel file that is just a list of employee numbers. He then wants his Disco query to only bring back the employee records for the people listed in the Excel file. Right now, he uses Cognos and can do this. A visual representation of this would be a condiition that would accomplish something like this:
    WHERE EMPLOYEE_NUMBER IN (Excel row 1, Excel row 2, Excel row 3, ...)
    We do not want to have to load the data into a table each time to accomplish this, but no other way readily jumps out at me. Anyone have any ideas?
    Thanks in advance,
    Jewell

    Jewell.
    Given the situation you describe, I don't know of any 'reasonable' way to do this either the GUI in Desktop or the interface in Plus.
    However - and just for the grins - there may be a possibility if done through the command line interface of Desktop. I truly doubt you'll want to do it - given the limitations - but ...
    ... in the command line interface you can put in the switch for /parameter such as: /parameter today trunc(SYSDATE), which means that your report is looking for a parameter called 'today' and you're putting in the data to put in the parameter 'trunc(SYSDATE)'.
    So, you could either build the command line with a batchy-kinda-process in that via batch it reads a file as input and puts parameters, or hard code it manually (which no one would want to do). Problem is, if you have multiple values, I don't think it'd work as you're saying something like:
    /parameter today (SYSDATE,SYSDATE-1,SYSDATE+!) - which I doubt would work as it probably only wants one value
    or if you had:
    /parameter today SYSDATE /parameter today SYSDATE-1 /parameter today SYSDATE+1
    it also might choke.
    Alternatively, you could investigate using Plus with the Java command line utility and see if it's possible to do a similar thing via Java.
    ... so long answer to a short question ... (my specialty) 8-)
    Russ

  • Local Cube Filtering / Add SSAS database as a data source for another SSAS project

    I need to generate a local cube using a fully processed SSAS database deployed on SSAS server. However, I need to restrict the data downloaded in the local cube while doing this.
    I have restricted access to the dimension data by mapping data to UserName/CustomData & using
    dynamic role-based security. When I generate a
    local cube by executing an AdoMdCommand using xmla structure of this online cube, the .cub file thus created also contains the data that shouldn't be accessible to the user creating the cube.
    I am aware of
    CREATE GLOBAL CUBE statement, but since it has some limitations such as we cannot have DISTINCT COUNT measures, I cannot use this statement because my online cube contains some important measures that use the DISTINCT COUNT aggregation. Plus, when I include
    a dimension with huge amount of data in it, the execution of command times out.
    So I am trying out an approach wherein I create a different SSAS project and use the existing SSAS database as a Data Source and then carry out the filtering to output the .cub file using xmla structure of this project database. The problem is, somehow SQL
    Server Data Tools 2012 does not recognize another SSAS database as a valid data source.
    Could someone please let me know why can't we do this or suggest me any alternative approach to implement this?
    Regards, &lt;br/&gt; Akshay

    Hi David,
    According to your description, you are creating a SQL Server Analysis Services project, now what you want is using SAP HANA as the data source, right?
    SSAS support many type of data source. However, as you can see on the link below, SAP HANA data source was not listed on that link. So this type of data source is not supported in current version of SSAS. Microsoft will update that document when it is supported.
    http://msdn.microsoft.com/en-IN/library/ms175608.aspx
    Thank you for your understanding.
    Regards,
    Charlie Liao
    TechNet Community Support

  • Handling parameters when using BI Server as a Data Source

    Hi,
    Following an upgrade of our BI Publisher 10g content to 11g I've come across a syntax issue with 11g when it comes to the handling of parameters in BI Server Data Sets.
    Under 10g, the documentation stated the following:
    +"If your data source is the Oracle BI Server, use the following macro to handle the null:+
    +{$ if ${sYear}='*'$}+
    +{$elsif ${sYear}='2000' $}+
    +where Year = :sYear+
    +{$else $}+
    +where Year = :sYear+
    +{$endif$} "+
    http://docs.oracle.com/cd/E10415_01/doc/bi.1013/e12187/T518230T518233.htm
    Under 11g it does not seem to allow the use of macros defined in Data Sets and raises the following error when saving the query:
    +"java.io.IOException: prepare query failed[nQSError: 43113] Message returned from OBIS. [nQSError: 27002] Near <$>: Syntax error [nQSError: 26012] ."+
    If macros are no longer supported, what is the best approach to handle null value parameters (caused by all choices / nothing being selected)
    as they need to be defined in the Data Set WHERE CLAUSE?
    I've tried a rather clunky CASE WHEN statement which handles string parameters fine, but dates are another thing as null values are returned as 'null' text which causes TYPE mismatch errors.
    Hopefully there is an elegant way to handle this in 11g (the 10g macros were very straight forward when dealing with BI Server)
    Many Thanks

    If what you are saying is true then I would highly recommend open a ticket with Oracle Support.
    They should provide and alternative for handling NULLS in 11g.
    regards
    Jorge

  • CR2008  Cannot report on IIS Log file data source

    I have CR2008 (SP0) and CRXIR2 installed on a Vista desktop.   I can create reports against IIS Log files using XIR2.   When I attempt to make the data connection using CR2008, I go through the same dialog to select log files and dates but at the end it displays "no items found" and I have no table connection that I can add to the report.
    My primary source is IIS 6 log files on a server using a mapped drive.  I have also tried the same using local IIS 7 on the same Vista pc that Crystal is installed on and neither work using CR2008;  both work fine from CRXIR2.
    I also opened from CR2008 an IIS log report created in CRXIR2 and oddly the report runs from 2008 but I cannot modify the data connection as I again just receive "no items found" if I attempt to establish another connection.
    I've tried a re-install and I've confirmed in Crystal setup that I have installed Web Activity Logs as Data source.

    Hi there,
    Try the following:
    1.  Connect to directory of where your IIS server that contains the log file.  It should be in C:\Windows\system32\LogFiles\W3SVC1
    (Or alternatively, I should suggest copying one of these files over locally to your workstation to test.  The file should be in the format of ex*.log)
    2.  Open up CR2008, and do a "Create New Connection"
    3.  Then choose "More Data Sources" -> "MS IIS/Proxy Log Files"
    4.  Point to where your logfile is, whether locally or remotely on the server. 
    5.  A "Select Log Files and Dates" window should appear
    6.  Under "Enter Log File Format and Location" panel (at the top), choose "Extend (ex*.log)" format
    7.  Browse to the file that you would to report off.

  • Data source in EP7

    Hi all gurus!
    We are trying to use Solution Manager as data source for logon in our EP7 but we can't get the settings right. We have used an existing installation as a "template" but it doesn't work and it might be because the other portal is of version 6.
    They are able to keep the config file dataSourceConfiguration_databese_only.xml and still see and use the tab "SAP System". Here they enter the data for their Solution Manager system and test it. They then go to config tool and change the file to dataSourceConfiguration_r3.xml, restart the server and everything work.
    To begin with, we can't se the tab "SAP System" at all but if we choose "ABAP System" in the drop down list "Data Source" (wich by the way doesn't contain a r3-alternative) a tab named "ABAP System" appears. Here we enter the data for Solution Manager but the test fails. So here is my first question:
    Is the template ABAP ok to use with SolMan?
    In the instructions from the other company, and which you saw above, it says that data source shall be database only and that you change to ...r3.xml in configtool. We tried that also, we chose "ABAP System", made the settings and then changed to ...r3.xml in configtool but the server won't start then. So, two more questions:
    Are the test fool proof? Since we disregarded the fact that the test failed and went on to configtool...
    Is the ...r3.xml a "valid" template to use in our version (was 7.0), since the server didn't start properly?
    And the last question:
    The user in SolMan that is used for the connection, of what type should he be? Service, Communication Data or something else or is without relevance? And what groups and roles should be added to him?
    Can anybody give some advice with this information?
    Best regards
    Benny

    The username isn't important, but if you add the SID at the end it means you can change the password for the SAPJSF user for a particular portal without affecting the other users.
    Anyway, to your problem.
    No tests are foolproof - if the test fails but the portal starts that is OK for me! When you select ABAP as the type, it will change automatiaclly to dataSourceConfiguration_abap.xml - leave this as is.
    Have you followed the instructions in <a href="http://help.sap.com/saphelp_nw04s/helpdata/en/49/9dd53f779c4e21e10000000a1550b0/frameset.htm">this link</a>? It is reasonably well documented.
    Cheers

Maybe you are looking for