Reports fail when run against a different data source

Hello,
We have a VB.NET 2008 WinForms application running on Microsoft .NET 3.5. We are using Crystal Reports 2008 runtime, service pack 3 -- using the CrystalDecisions.Windows.Forms.CrystalReportViewer in the app to view reports. In the GAC on all our client computers, we have versions 12.0.1100.0 and 12.0.2000.0 of CrystalDecisions.CrystalReports.Engine, CrystalDecisions.Shared, and CrystalDecisions.Windows.Forms.
Please refer to another one of our posted forum issues, u201CCritical issue since upgrading from CR9 to CR2008u201D, as these issues seem to be related:
Critical issue since upgrading from CR9 to CR2008
We were concerned with report display slow down, and we seemed to have solved this by using the Oracle Server driver (instead of either Microsoft's or Oracle's OLEDB driver).  But now we must find a resolution to another piece of the puzzle, which is:  why does a report break if one data source is embedded in the .rpt file is different than the one you are trying to run the report against, in the .NET Viewer?
Problem:
If you have a production database name (e.g. "ProdDB") embedded in your .rpt file that you built your report from and try to run that report against a development database (e.g. "DevDB") (OR VICE VERSA -- it is the switch that is the important concept here), the report fails with a list of messages such as this:
    Failed to retrieve data from the database
    Details:  [Database vendor code: 6550 ]
This only seems to happen if the source of the report data (i.e. the underlying query) is an Oracle stored procedure or a Crystal Reports SQL Command -- the reports run fine against all data sources if the source is a table or a view).  In trying different things to troubleshoot this, including adding a ReportDocument.VerifyDatabase() call after setting the connection information, the Crystal Reports viewer will spit out other nonsensical errers regarding being unable to find certain fields (e.g. "The field name is not known), or not able to find the table (even though the source data should be coming from an Oracle stored procedure, not a table).
When the reports are run in the Crystal Reports Designer, they run fine no matter what database is being used; but the problem only happens while being run in the .NET viewer.  It's almost as if something internally isn't getting fully "set" to the new data source, or something -- we're really grasping at straws here.
For the sake of completeness of information, here is how we're setting the connection information
        '-- Set database connection info for the main report
        For Each oConnectionInfo In oCrystalReport.DataSourceConnections
            oConnectionInfo.SetConnection(gsDBDataSource, "", gsDBUserID, gsDBPassword)
        Next oConnectionInfo
        '-- Set database connection info for each subreport
        For Each oSubreport In oCrystalReport.Subreports
            For Each oConnectionInfo In oSubreport.DataSourceConnections
                oConnectionInfo.SetConnection(gsDBDataSource, "", gsDBUserID, gsDBPassword)
            Next oConnectionInfo
        Next oSubreport
... but in troubleshooting, we've even tried an "overkill" approach and added this code as well:
        '-- Set database connection info for each table in the main report
        For Each oTable In oCrystalReport.Database.Tables
            With oTable.LogOnInfo.ConnectionInfo
                .ServerName = gsDBDataSource
                .UserID = gsDBUserID
                .Password = gsDBPassword
                For Each oPair In .LogonProperties
                    If UCase(CStr(oPair.Name)) = "DATA SOURCE" Then
                        oPair.Value = gsDBDataSource
                        Exit For
                    End If
                Next oPair
            End With
            oTable.ApplyLogOnInfo(oTable.LogOnInfo)
        Next oTable
        '-- Set database connection info for each table in each subreport
        For Each oSubreport In oCrystalReport.Subreports
            For Each oTable In oSubreport.Database.Tables
                With oTable.LogOnInfo.ConnectionInfo
                    .ServerName = gsDBDataSource
                    .UserID = gsDBUserID
                    .Password = gsDBPassword
                    For Each oPair In .LogonProperties
                        If UCase(CStr(oPair.Name)) = "DATA SOURCE" Then
                            oPair.Value = gsDBDataSource
                            Exit For
                        End If
                    Next oPair
                End With
                oTable.ApplyLogOnInfo(oTable.LogOnInfo)
            Next oTable
        Next oSubreport
... alas, it makes no difference.  If we run the report against a database that is different than the one specified with "Set Datasource Location" in Crystal, it fails with nonsense errors 

Thanks for the reply, Ludek.  We have made some breakthroughs, uncovered some Crystal bugs and workarounds, and we're probably 90% there I hope.
For your first point, unfortunately the information on the Oracle 6550 error was generic, and not much help in our case.  And for your second point, the errors didn't have anything to do with subreports at that time -- the error would manifest itself even in a simple, one-level report.
However, your third point (pointing us to KB 1553921) helped move us forward quite a bit more.  For the benefit of all, here is a link to that KB article:
Link: [KB 1553921|http://www.sdn.sap.com/irj/servlet/prt/portal/prtroot/com.sap.km.cm.docs/oss_notes_boj/sdn_oss_boj_bi/sap(bD1lbiZjPTAwMQ==)/bc/bsp/spn/scn_bosap/notes%7B6163636573733d36393736354636443646363436353344333933393338323636393736354637333631373036453646373436353733354636453735364436323635373233443330333033303331333533353333333933323331%7D.do]
We downloaded the tool referenced there, and pointed it at a couple of our reports.  The bottom line is that the code it generated uses a completely new area of the Crystal Reports .NET API which we had not used before -- in the CrystalDecisions.ReportAppServer namespace.  Using code based on what that RasConnectionInfo tool generated, we were able gain greater visibility into some of the objects in the API and to uncover what I think qualifies as a genuine bug in Crystal Reports.
The CrystalDecisions.ReportAppServer.DataDefModel.ISCRTable class exposes a property called QualifiedName, something that isn't exposed by the more commonly-used CrystalDecisions.CrystalReports.Engine.Table class.  When changing the data source with our old code referenced above (CrystalDecisions.Shared.ConnectionInfo.SetConnection), I saw that Crystal would actually change the Table.QualifiedName from something like "SCHEMAOWNER.PACKAGENAME.PROCNAME" to just "PROCNAME" (essentially stripping off the schema and package name).  Bad, Crystal...  VERY BAD!  IMHO, Crystal potentially deserves to be swatted on the a** with the proverbial rolled-up newspaper.
I believe this explains why we were also able to generate errors indicating that field names or tables were not found -- because Crystal had gone and changed the QualifiedName to remove some key info identifying the database object!  So, knowing this and using the code generated by the RasConnectionInfo tool, we were able to work around this bug with code that worked for most of our reports ("most" is the key word here -- more on that in a bit).
So, first of all, I'll post our new code.  Here is the main area where we loop through all of the tables in the report and subreports:
'-- Replace each table in the main report with new connection info
For Each oTable In oCrystalReport.ReportClientDocument.DatabaseController.Database.Tables
    oNewTable = oTable.Clone()
    oNewTable.ConnectionInfo = GetNewConnectionInfo(oTable)
    oCrystalReport.ReportClientDocument.DatabaseController.SetTableLocation(oTable, oNewTable)
Next oTable
'-- Replace each table in any subreports with new connection info
For iLoop = 0 To oCrystalReport.Subreports.Count - 1
    sSubreportName = oCrystalReport.Subreports(iLoop).Name
    For Each oTable In oCrystalReport.ReportClientDocument.SubreportController.GetSubreportDatabase(sSubreportName).Tables
        oNewTable = oTable.Clone()
        oNewTable.ConnectionInfo = GetNewConnectionInfo(oTable)
        oCrystalReport.ReportClientDocument.SubreportController.SetTableLocation(sSubreportName, oTable, oNewTable)
    Next oTable
Next iLoop
'-- Call VerifyDatabase() to ensure that the tables update properly
oCrystalReport.VerifyDatabase()
(Thanks to Colin Stynes for his post in the following thread, which describes how to handle the subreports):
Setting subreport connection info at runtime
There seems to be a limitation on the number of characters in a post on this forum (before all formatting gets lost), so please see my next post for the rest....

Similar Messages

  • Reports fail when running across a WAN connection

    We have a remote site in Detroit that has excutables installed on local file server but connects to a database here in Memphis.  Since doing this, we are having issues running some reports from Detroit.  When we attempt to run the reports, we get an error "Failed to retrieve data from the database. Detail:  [Database Vendor Code:11]".  The database is SQL 2005.  The following facts are true:
    1. Certain reports will not run from production PCs in Detroit; these PC just have a Crytal 11.5 runtime installed.
    2. We can run the exact same report (same .rpt file) from production PCs in Memphis using the same paramters and it runs fine.  The PCs are configure the same as the ones in Detroit.
    3. We can run the exact same report from a development PC in Detroit (has Crystal 11.5 installed) and it runs fine.
    Since some reports do run, it seems that the workstation is configured properly, and since the reports do run with no errors from PCs in Memphis, it seems that there is nothing wrong with the .rpt files, the data, or the database.  So, it seems to boil down to something with the WAN connection possibly.  I'm out of ideas at this point.  This issue really seems to make no sense, and any help would be appreciated.

    Hi Clint,
    You should have a link to log a Message in Service Market Place ( SMP ). From there once the case is enter an Engineer will pick it up and give you a call to discuss.
    313 is the original release and you should upgrade to at least SP4.
    What you should do is install the SP4 update to your test machine first and then download and deploy the runtime using the same version of merge modules or MSI. Updating runtime without re-building your app is likely going cause issues. You should also test all this on a test machine first for the obvious reasons.
    SP4 Merge modules are available, not available for SP 5 yet, they are working on getting them posted for download.
    You did not say what technology ( report engine ) you are using, RDC or CR .NET assemblies?
    It may also be due to different MDAC or MS SQL Server client installed. 2005 has a separate Client install now, previously MDAC was used to install the client, depends on which connection type you are using.
    It may also be a table or database permission issue. MS SQL Server enhanced their security and fully support DB security. Try a different user to see if that works.
    Also to clarify, a UDL file does NOT use ODBC unless you select Microsofts OLE DB for ODBC. Create a txt file and then rename the extension to udl. Double click the file and the OLE DB UI should pop up, then select MS SQL Server OLE DB Provider. Fill in the connection info and test it. This may explain why it's failing also.
    Thank you
    Don
    Edited by: Don Williams on Apr 1, 2009 10:30 AM

  • Report fails when run in Workspace, but it works in  reporting studio

    Hi all,
    I have a bqy file which runs fine in the design studio client (v9.3) however when I import it into the Workspace and try and run it i get the following error :
    An Interactive Reporting Service error has occurred.-SQL API: [SQLFetchScroll], SQL RETURN: [-1], SQL STATE: [HY010], SQL NATIVE ERROR: [0], SQL MESSAGE: [[DataDirect][ODBC lib] Function sequence error]
    (2001)
    The query was built from importing the SQL as below:
    DECLARE @Fig1 int
    DECLARE @i int
    DECLARE skuCursor CURSOR FOR
    Select r.QtyReceived / b.qty as Figure1
    from ReceiptDetail r
    inner join BillofMaterial b on b.SKU = r.Lottable03
    where r.DateReceived > getdate() - 14
    create table #TEMPIG (Test char(2),Median1 int)
    OPEN skuCursor
    FETCH NEXT FROM skuCursor INTO @Fig1
    set @i = 0
    WHILE @@FETCH_STATUS = 0
    BEGIN
    while @i < @Fig1
    begin
    insert into #TempIG (Test,Median1) values ('MV',@Fig1)
    --print @Fig1
    set @i = @i + 1
    end
    set @i = 0
    FETCH NEXT FROM skuCursor INTO @Fig1
    END
    CLOSE skuCursor
    DEALLOCATE skuCursor
    --select cast(Median1 as Int) from #TEMPIG
    --order by 1 asc
    select
    AVG(Median1) as MedianVal
    from
    select Median1,
    ROW_NUMBER() over (partition by Test order by Median1 ASC) as MedRank,
    COUNT(*) over (partition by Test) as MedCount
    from
    #TEMPIG
    ) x
    where
    x.MedRank in (x.MedCount/2+1, (x.MedCount+1)/2)
    drop table #TEMPIG
    Could anyone explain what is wrong and how to put it right.
    Many Thanks in advance.
    Ian.

    Hi,
    OCEs the same? Yes
    Client drivers the same? Yes
    DAS connection information set up same as desktop connection information Yes
    The server which workspace runs is on Linux, but the reporting studio client runs on windows desktops, this is the only difference i know off.
    I have chopped and changed things whilst testing this report and it appears that the fault relates to the Cursor and the navigation (Fetch) through the records.
    This is starting to drive me around the bend.
    Please do come back to me if you have any ideas.
    Many thanks in advance.
    Ian.

  • Wfsdupld fails when run against 8.1.7 database on 64 bit Sun SPARC Solaris 8

    Attempting to run in the seed data.
    Platform is 64 bit Solaris v8
    Database is 8.1.7.0
    Workflow 2.6
    Everything is fine until we run the wfsdupld script. The script
    fails with
    ORA-29516 Aurora Assertion Failed: Assertion failure at
    joncomp.c:127
    jtc_active_clint_ncomp_slots (oracle/xml/parser/v2/DTD, 0)
    returned 0
    ORA-6512 at OWF_MGR.WF_EVENT_SYNCHRONIZE_PKG line 373
    Any ideas?

    I raised this as a TAR #1838186.995 and received the following
    workaround which does seem to have worked on 8.1.7.0 >>
    I have taken a look at bug 2034596 which suggests a problem with
    the 8.1.7.0 database. However, by upgrading the database to
    8.1.7.2 the error was replaced by a nullpointerexception.
    As a possible workaround it may be worth trying the following
    (on a 8.1.7.0 database):
    $ svrmgrl
    SVRMGR>connect internal
    SVRMGR>alter system flush shared_pool;
    SVRMGR>alter system flush shared_pool;
    SVRMGR>alter system flush shared_pool;
    SVRMGR>alter java class "java/io/Serializable" check;
    SVRMGR>alter system flush shared_pool;
    SVRMGR>alter system flush shared_pool;
    SVRMGR>alter system flush shared_pool;
    SVRMGR>shutdown immediate;
    SVRMGR> startup

  • Login Module validation against two different data sources

    Hi All,
    We have the following setup
    NW Portal 7.31
    Single stack
    We are in the process of implementing Employee Self Service 1.50 WDA.
    And as part of the requirement , we have implemented LDAP as the UME DataSource.
    So, employees log into the portal using the LAN ID and Windows password  and are authenticated using the Logon ticket.
    Now the client has come up with a requirement to have all German employees login with their ECC user id and password.
    NOTE: ECC User ID and LAN IDs are the same, but the passwords are different.
    a) Is it possible to use two UME data sources for logon validation.
    b) Is customization of the Login module needed for this scenario?? If Yes, what components need to customized?
    c) What is the best approach to handle this requirement?
    Thanks.
    Regards,
    Raj

    check this link if it provide any helpful info:
    User Mappings in the Authentication Framework of SAP NetWeaver Application Server (AS) Java - Security and Identity Mana…
    Also you can search around creating a custom login module for authenticating useres against ECC ,if its possible you need to plugin the custom login module in your authentication stack along with the applicable flag and order so that if user authentication failed at ldap level then the control should go to authenticate it against the ECCcustom login module and logon ticket should be generated in both the cases.

  • Error when running a CO-PA data source

    Hi all, I just created a CO-PA data source,
    when I try to run it in RSA6 Im getting the following error:
    No source could be found to answer the search query
    Any body knows what might be happening?
    pleaseeeee help.
    thanks

    Hi, it is available, the problem is when I try to run it, then I get that error.
    I went to transaction KEDV and all the camps in the summarization level are in my data source, so then I dont know what can be the problem,  any ideas?

  • Sdo_filter fail when query against a spatial view in different schema

    We have a table with X,Y coordinates and would like to run spatial query against it. We do not want to change the table structure, so we opt to use a function based index. USER_SDO_GEOM_METADATA is updated and index is built. Then we created a view with spatial column from the table. Everything works fine with the user who owns the table and view.
    When we try to run a spatial query against the view from a different user, it failed with error. However, if we substitute the select from my_view* with the actual SQL statement that created the view, it works. So it looks like Oracle refuse to acknowledge the spatial index if accessed via view. Here is some simplified scripts:
    --- connect as USER1.
    --update meta data
    INSERT INTO USER_SDO_GEOM_METADATA ( TABLE_NAME, COLUMN_NAME, DIMINFO, SRID ) VALUES
    ('LOCATIONS', 'MDSYS.SDO_GEOMETRY(2001,2264,SDO_POINT_TYPE(NVL(X_COORD,0),NVL(Y_COORD,0),NULL),NULL,NULL)',
    SDO_DIM_ARRAY( SDO_DIM_ELEMENT('X', 1300000, 1600000, 1), SDO_DIM_ELEMENT('Y', 400000, 700000, 1) ), 2264 );
    --created index
    CREATE INDEX LOCA_XYGEOM_IDX ON LOCATIONS
    ( SDO_GEOMETRY(2001,2264,SDO_POINT_TYPE(NVL(X_COORD,0),NVL(Y_COORD,0),NULL),NULL,NULL)
    ) INDEXTYPE IS MDSYS.SPATIAL_INDEX;
    --create view
    CREATE VIEW USER1.MY_VIEW AS SELECT ID ,X_COORD,Y_COORD, SDO_GEOMETRY(2001,2264,SDO_POINT_TYPE(NVL(X_COORD,0),NVL(Y_COORD,0),NULL),NULL,NULL) SHAPE
    FROM USER1.LOCATIONS WHERE X_COORD>0 AND Y_COORD>0;
    -- run spatial query from view, works fine by user1 by failed on user2.
    SELECT SHAPE FROM (
    SELECT * FROM USER1.MY_VIEW
    ) a WHERE sdo_filter (shape, sdo_geometry ('POLYGON ((1447000 540000, 1453000 540000, 1453000 545000, 1447000 545000, 1447000 540000))', 2264), 'querytype=window') = 'TRUE';
    -- run spatial query from table directly, simply replace the view with actual statements that created the view. works fine by user1 AND user2.
    SELECT SHAPE FROM (
    SELECT ID ,X_COORD,Y_COORD, SDO_GEOMETRY(2001,2264,SDO_POINT_TYPE(NVL(X_COORD,0),NVL(Y_COORD,0),NULL),NULL,NULL) SHAPE
    FROM USER1.LOCATIONS WHERE X_COORD>0 AND Y_COORD>0
    ) a WHERE sdo_filter (shape, sdo_geometry ('POLYGON ((1447000 540000, 1453000 540000, 1453000 545000, 1447000 545000, 1447000 540000))', 2264), 'querytype=window') = 'TRUE';
    When run against the view by user2, the error is:
    ORA-13226: interface not supported without a spatial index
    ORA-06512: at "MDSYS.MD", line 1723
    ORA-06512: at "MDSYS.MDERR", line 8
    ORA-06512: at "MDSYS.SDO_3GL", line 1173
    13226. 00000 - "interface not supported without a spatial index"
    *Cause:    The geometry table does not have a spatial index.
    *Action:   Verify that the geometry table referenced in the spatial operator
    has a spatial index on it.
    Note, the SELECT SHAPE FROM (****) A WHERE SDO_FILTER(....) syntax is a third party application, all we have control is the part inside "(select ...)".
    So it appears Oracle is treating view differently. Have attempted fake the view name into USER_SDO_GEOM_METADATA, did not work. Also, granted select on the index table to user2, did not work.
    if we re-created the view in user2 schema, it worked for user2 but not user1, so it's not something we can do for every user.
    Searched the forum, no good match found. A few posts talked about "union all" in view caused the problem but I do not have the union.
    We are only use Oracle 10g Locator, not full spatial edition.
    Any ideas?
    Thanks!
    Edited by: liu.284 on Oct 4, 2011 12:08 PM

    It seems a bug, where a function-based spatial index is not correctly handled in a view query transformation.
    Not sure if the following works for you or not.
    add a new column "shape" (mdsys.sdo_geometry) in table locations, use a trigger and x_coord/y_coord
    to set values for this new column, and just create a normal spatial index on this new column. (drop the
    function-based spatial index). And create a view like:
    CREATE VIEW USER1.MY_VIEW2 AS SELECT ID , X_COORD, Y_COORD, SHAPE
    FROM USER1.LOCATIONS WHERE X_COORD>0 AND Y_COORD>0;

  • I have a production mobile Flex app that uses RemoteObject calls for all data access, and it's working well, except for a new remote call I just added that only fails when running with a release build.  The same call works fine when running on the device

    I have a production mobile Flex app that uses RemoteObject calls for all data access, and it's working well, except for a new remote call I just added that only fails when running with a release build. The same call works fine when running on the device (iPhone) using debug build. When running with a release build, the result handler is never called (nor is the fault handler called). Viewing the BlazeDS logs in debug mode, the call is received and send back with data. I've narrowed it down to what seems to be a data size issue.
    I have targeted one specific data call that returns in the String value a string length of 44kb, which fails in the release build (result or fault handler never called), but the result handler is called as expected in debug build. When I do not populate the String value (in server side Java code) on the object (just set it empty string), the result handler is then called, and the object is returned (release build).
    The custom object being returned in the call is a very a simple object, with getters/setters for simple types boolean, int, String, and one org.23c.dom.Document type. This same object type is used on other other RemoteObject calls (different data) and works fine (release and debug builds). I originally was returning as a Document, but, just to make sure this wasn't the problem, changed the value to be returned to a String, just to rule out XML/Dom issues in serialization.
    I don't understand 1) why the release build vs. debug build behavior is different for a RemoteObject call, 2) why the calls work in debug build when sending over a somewhat large (but, not unreasonable) amount of data in a String object, but not in release build.
    I have't tried to find out exactly where the failure point in size is, but, not sure that's even relevant, since 44kb isn't an unreasonable size to expect.
    By turning on the Debug mode in BlazeDS, I can see the object and it's attributes being serialized and everything looks good there. The calls are received and processed appropriately in BlazeDS for both debug and release build testing.
    Anyone have an idea on other things to try to debug/resolve this?
    Platform testing is BlazeDS 4, Flashbuilder 4.7, Websphere 8 server, iPhone (iOS 7.1.2). Tried using multiple Flex SDK's 4.12 to the latest 4.13, with no change in behavior.
    Thanks!

    After a week's worth of debugging, I found the issue.
    The Java type returned from the call was defined as ArrayList.  Changing it to List resolved the problem.
    I'm not sure why ArrayList isn't a valid return type, I've been looking at the Adobe docs, and still can't see why this isn't valid.  And, why it works in Debug mode and not in Release build is even stranger.  Maybe someone can shed some light on the logic here to me.

  • Not able to edit the report created on different data source.....

    I have a query regarding Report in OBIEE - reports developed from BI Publisher are specific to data source on which they have been created ??
    i have a sample report that was created on different data source, i have the corresponding RPD also. I changed the data source according to my DB and when i try to update/edit the report,
    on Analytics for adding a new column, it is generating a seperate new Query from QueryBuilder for that additional cloumn rather than adding up the new query with the previous one(existing report query). Is it because of mismatch of data source on which report had been created and on which it is being update ?? if it is the case, where do i need to make changes related to JDBC connection or others ??
    when i try to create a new sample data set and try to update it, it adds up the extra edited things to original query and works perfectly fine. can ny 1 help me for the same ??

    Hi Denis,
    Normally,what we do is once we provide access to webi users group for each user from BO supervisor module, user(s) will able to refresh/edit the existing report from Full client BO.His colleagues have had no problem editing all his reports from their machines but he is not able to edit any report from his machine and BO is getting freeze.
    He also reinstalled BO and cleaned out everything on his Computer and re-built it but the issue is not yet resolved
    Can you please tell me how to resolve this issue
    Kind Regards,
    Srinivas

  • Report using two different data sources won't work.

    I'm trying to build a report that shows information from a production table and an archive table.
    Tables are in different databases, which are defined as their own Data Sources in Publisher.
    Two data sets containing the same query but using different Data Sources are defined in the Data Model.
    When selecting option 'Concatenated SQL Data Source' the report never completes.
    If any of the two Data Sets is selected as the Main Data Set, the report shows information related to that source only.
    Any hints on how to make this work would be appreciated.
    Thanks.
    ccastillo

    More details on this issue:
    The production database has a synonym pointing to the archive database. I build a query using a UNION ALL statement linking both tables.
    For the same set of parameters, this query completes in a couple of minutes outside BI Publisher, but never ends (I cancel after an hour) inside Publisher.
    Is there any special considerations for the use of synonyms inside Publisher?

  • ORA-04062 error when running forms with different users

    ORA-04062 error when running forms with different users
    I have a form that has a block that should display some data from another users tables. (The other user's name is dynamic, it's selected from a list box)
    I wrote a stored procedure to get the data from other user's tables.
    When I compile the form and run it with the same user I compiled, it works without any error. But when I run the compiled form with another user I get the ORA-04062 (signature of procedure has been changed) error.
    I tried setting REMOTE_DEPENDENCIES_MODE to SIGNATURE in init.ora but it didn't help.
    My Forms version is 6i with Patch 15.
    Database version is 9.
    Here is my stored procedure:
    TYPE Scenario_Tab IS TABLE OF NUMBER(34) INDEX BY BINARY INTEGER;
    TYPE Open_Curs IS REF CURSOR;
    PROCEDURE Get_Scenarios(User_Name IN VARCHAR2, Scen_Table OUT Scenario_Tab) IS
    Curs Open_Curs;
    i NUMBER;
    BEGIN
    OPEN Curs FOR
    'SELECT Seq_No FROM '|| User_Name ||'.scenario';
    i := 1;
    LOOP
    FETCH Curs INTO Scen_Table(i);
    EXIT WHEN Curs%NOTFOUND;
    i := i + 1;
    END LOOP;
    END Get_Senarios;
    I would be happy to solve this problem. It's really important.
    Maybe somebody can tell me another way to do what I want to do. (getting a list of values from another users tables)

    I think it should be a better solution to create a package,
    and put your own TYPES and procedure into it.
    CREATE OR REPLACE PACKAGE PKG_XXX IS
    TYPE TYP_TAB_CHAR IS TABLE OF .... ;
    PROCEDURE P_XX ( Var1 IN VARCHAR2, var2 IN OUT TYP_TAB_CHAR );
    END ;
    Then in your Form :
    Declare
    var PKG_XXX.TYP_TAB_CHAR ;
    Begin
    PKG_XXX.P_XX( 'user_name', var ) ;
    End ;

  • 11gR2 RAC install fail when running root.sh script on second node

    I get the errors:
    ORA-15018: diskgroup cannot be created
    ORA-15072: command requires at least 2 regular failure groups, discovered only 0
    ORA-15080: synchronous I/O operation to a disk failed
    [main] [ 2012-04-10 16:44:12.564 EDT ] [UsmcaLogger.logException:175] oracle.sysman.assistants.util.sqlEngine.SQLFatalErrorException: ORA-15018: diskgroup cannot be created
    ORA-15072: command requires at least 2 regular failure groups, discovered only 0
    ORA-15080: synchronous I/O operation to a disk failed
    I have tried the fix solutions from metalink note, but did not fix issue
    11GR2 GRID INFRASTRUCTURE INSTALLATION FAILS WHEN RUNNING ROOT.SH ON NODE 2 OF RAC USING ASMLIB [ID 1059847.1                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                   

    Hi,
    it looks like, that your "shared device" you are using is not really shared.
    The second node does "create an ASM diskgroup" and create OCR and Voting disks. If this indeed would be a shared device, he should have recognized, that your disk is shared.
    So as a result your VMware configuration must be wrong, and the disk you presented as shared disk is not really shared.
    Which VMWare version did you use? It will not work correctly with the workstation or player edition, since shared disks are only really working with the server version.
    If you indeed using the server, could you paste your vm configurations?
    Furthermore I recommend using Virtual Box. There is a nice how-to:
    http://www.oracle-base.com/articles/11g/OracleDB11gR2RACInstallationOnOEL5UsingVirtualBox.php
    Sebastian

  • LSMW Fails when run in B/G but works fine in Front end..why?

    Hi All,
    i am trying to run a batch process by LSMW, my files are accurate, no problem with them, everything works fine but it fails when run in BG..works absolutely fine in front end. whats the diff with running in B/G?
    same thing happens when i am trying to execute an RFC thru SAP JCO, it works when debugger is on (i guess switching on debugger is similar to running in B/G) but it doesnt work when debugger is off. but when i execute that RFC directly in se37 from SAP gui it works fine..fails when connected to JCO..
    i am not having this issue with r/3 4.6c or mySAP ECC.6.0  i have this issue only in r/3 4.7.
    has anyone faced the similar situtaion? pls help.
    thanks.
    p.s if this may help. the RFC and LSMW error both are pertaining to change in address of US employees.( infotype 0006)

    Applying  SAP note 928273 Solved this issue.

  • LSMW fails when run in B/G works fine in Frontend..why?

    Hi All,
    i am trying to runa batch process by LSMW, my files are accurate, no problem with them, everything works fine but it fails when run in BG..works absolutely fine in front end. whats the diff with running in B/G?
    same thing happens when i am trying to execute an RFC thru SAP JCO, it works when debugger is on 9i guess switching on debugger is similar to running in B/G) but it doesnt work when debugger is off. but when i execute that RFC directly in se37 from SAP gui it works fine..fails when connected to JCO..
    i am not having this issue with r/3 4.6c or mySAP ECC.6.0  i have this issue only in r/3 4.7.
    has anyone faced the similar situtaion? pls help.
    thanks.
    p.s if this may help. the RFC and LSMW error both are pertaining to change in address of US employees.( infotype 0006)

    for LSMW its the recording of transaction PA40 (employee hire fails when filling  address details) and PA30 (change address) and same is the case with RFC..well its a BAPI_ADDRESSEMPUS_CHANGE. 
    To eloborate more..the error is..Fill in all the mandatory fields.
    which i am very much doing..there are no hidden fields or anything..i have seen the screens etc..I AM filling all mandatory fields. infact i am not leaving anything unfilled., same scrren is going fine when in front end..i am just clicking ok..ok..ok and boom transaction complete..no complaints. but running B/G is killing me.
    i have to run batch for 100,000 employees
    What fails my logic is..its working fine in 4.6c and mySAP ECC.6.0 but not in 4.7
            Hruser
    Message was edited by:
            Hruser

  • LSMW fails when run in B/G but works in Frontend..why?

    Hi All,
    i am trying to runa batch process by LSMW, my files are accurate, no problem with them, everything works fine but it fails when run in BG..works absolutely fine in front end. whats the diff with running in B/G?
    same thing happens when i am trying to execute an RFC thru SAP JCO, it works when debugger is on 9i guess switching on debugger is similar to running in B/G) but it doesnt work when debugger is off. but when i execute that RFC directly in se37 from SAP gui it works fine..fails when connected to JCO..
    i am not having this issue with r/3 4.6c or mySAP ECC.6.0  i have this issue only in r/3 4.7.
    has anyone faced the similar situtaion? pls help.
    thanks.
    p.s if this may help. the RFC and LSMW error both are pertaining to change in address of US employees.( infotype 0006)

    Applying SAP note 928273 Solved this issue.
    thank you.

Maybe you are looking for

  • Push and birthday calendar issues

    I'm having trouble with push. As a somewhat separate issue, I'm also having trouble with birthdays from iCal to the iPhone, push or not. I had no problem with birthdays transferring to my old iPhone. Should I set up MobileMe as .mac or .me, since I h

  • Need to Install 2 new HDD in Storedge 3320 (Raid 0+1)

    Hello, I have ST3320 connected to V440 Actual configuration of ST3320= 3 x 3 HDD / RAID 0+1 I have to ad 2 new disks and keep the RAID 0+1 How should I proceed? PS. * Network interface: YES ip-address: 192.168.1.1 netmask: 255.255.255.0 gateway: 0.0.

  • Call Minder Notifications never arrive on Orange M...

    Call Minder notifications of waiting messages on the home phone fail to arrive on my Orange Mobile phone.  Orange claim it is a BT problem .  When I redirect my notifications to another mobile carrier the notifications arrive almost immediately.  Hav

  • Smallest file size for talking head, MPEG1?

    Hey friends, back again, still using PRE8. I have to encode a talking head into a watchable quality for a podcast. Prefer between 320x240 and 640x480, prefer the later. I inadvertently tried some combination of VCD 352x240 and got an amazingly small

  • HTML a tags in report

    Hi Guys, I have an updateable report in which the SQL calls a PLSQL function which adds HTML <a> tags to the returned value. This hyperlink is then used to call a seperate web page. Since our upgrade to v4, the report is acutually displaying the text