Login Module validation against two different data sources

Hi All,
We have the following setup
NW Portal 7.31
Single stack
We are in the process of implementing Employee Self Service 1.50 WDA.
And as part of the requirement , we have implemented LDAP as the UME DataSource.
So, employees log into the portal using the LAN ID and Windows password  and are authenticated using the Logon ticket.
Now the client has come up with a requirement to have all German employees login with their ECC user id and password.
NOTE: ECC User ID and LAN IDs are the same, but the passwords are different.
a) Is it possible to use two UME data sources for logon validation.
b) Is customization of the Login module needed for this scenario?? If Yes, what components need to customized?
c) What is the best approach to handle this requirement?
Thanks.
Regards,
Raj

check this link if it provide any helpful info:
User Mappings in the Authentication Framework of SAP NetWeaver Application Server (AS) Java - Security and Identity Mana…
Also you can search around creating a custom login module for authenticating useres against ECC ,if its possible you need to plugin the custom login module in your authentication stack along with the applicable flag and order so that if user authentication failed at ldap level then the control should go to authenticate it against the ECCcustom login module and logon ticket should be generated in both the cases.

Similar Messages

  • Report using two different data sources won't work.

    I'm trying to build a report that shows information from a production table and an archive table.
    Tables are in different databases, which are defined as their own Data Sources in Publisher.
    Two data sets containing the same query but using different Data Sources are defined in the Data Model.
    When selecting option 'Concatenated SQL Data Source' the report never completes.
    If any of the two Data Sets is selected as the Main Data Set, the report shows information related to that source only.
    Any hints on how to make this work would be appreciated.
    Thanks.
    ccastillo

    More details on this issue:
    The production database has a synonym pointing to the archive database. I build a query using a UNION ALL statement linking both tables.
    For the same set of parameters, this query completes in a couple of minutes outside BI Publisher, but never ends (I cancel after an hour) inside Publisher.
    Is there any special considerations for the use of synonyms inside Publisher?

  • Two Different Data Sources in EIS

    <p>Hi -</p><p> </p><p>Can we have 2 different Data Sources in the OLAP model.</p><p>Example:</p><p>We have our Fact table in Teradata and Dimension table inOracle.</p><p>Can we create a model with this kind of scenario?</p><p>WIll it be a maintenance nightmare if we do this?</p><p> </p><p>Any help would be greatly appreciated.</p><p> </p><p>Thanks,</p><p>Jeeva</p>

    Hi Jeeva,<BR><BR>Yes you can do this. I have models which use data from Oracle, SQL Server and also Text files! All you need is to setup the appropriate ODBC connections on the box that contains EIS then when you create your OLAP model, you add new data sources by clicking "Connections" then "Add Data Source"<BR><BR>Hope this helps.<BR><BR>Brian

  • Reports fail when run against a different data source

    Hello,
    We have a VB.NET 2008 WinForms application running on Microsoft .NET 3.5. We are using Crystal Reports 2008 runtime, service pack 3 -- using the CrystalDecisions.Windows.Forms.CrystalReportViewer in the app to view reports. In the GAC on all our client computers, we have versions 12.0.1100.0 and 12.0.2000.0 of CrystalDecisions.CrystalReports.Engine, CrystalDecisions.Shared, and CrystalDecisions.Windows.Forms.
    Please refer to another one of our posted forum issues, u201CCritical issue since upgrading from CR9 to CR2008u201D, as these issues seem to be related:
    Critical issue since upgrading from CR9 to CR2008
    We were concerned with report display slow down, and we seemed to have solved this by using the Oracle Server driver (instead of either Microsoft's or Oracle's OLEDB driver).  But now we must find a resolution to another piece of the puzzle, which is:  why does a report break if one data source is embedded in the .rpt file is different than the one you are trying to run the report against, in the .NET Viewer?
    Problem:
    If you have a production database name (e.g. "ProdDB") embedded in your .rpt file that you built your report from and try to run that report against a development database (e.g. "DevDB") (OR VICE VERSA -- it is the switch that is the important concept here), the report fails with a list of messages such as this:
        Failed to retrieve data from the database
        Details:  [Database vendor code: 6550 ]
    This only seems to happen if the source of the report data (i.e. the underlying query) is an Oracle stored procedure or a Crystal Reports SQL Command -- the reports run fine against all data sources if the source is a table or a view).  In trying different things to troubleshoot this, including adding a ReportDocument.VerifyDatabase() call after setting the connection information, the Crystal Reports viewer will spit out other nonsensical errers regarding being unable to find certain fields (e.g. "The field name is not known), or not able to find the table (even though the source data should be coming from an Oracle stored procedure, not a table).
    When the reports are run in the Crystal Reports Designer, they run fine no matter what database is being used; but the problem only happens while being run in the .NET viewer.  It's almost as if something internally isn't getting fully "set" to the new data source, or something -- we're really grasping at straws here.
    For the sake of completeness of information, here is how we're setting the connection information
            '-- Set database connection info for the main report
            For Each oConnectionInfo In oCrystalReport.DataSourceConnections
                oConnectionInfo.SetConnection(gsDBDataSource, "", gsDBUserID, gsDBPassword)
            Next oConnectionInfo
            '-- Set database connection info for each subreport
            For Each oSubreport In oCrystalReport.Subreports
                For Each oConnectionInfo In oSubreport.DataSourceConnections
                    oConnectionInfo.SetConnection(gsDBDataSource, "", gsDBUserID, gsDBPassword)
                Next oConnectionInfo
            Next oSubreport
    ... but in troubleshooting, we've even tried an "overkill" approach and added this code as well:
            '-- Set database connection info for each table in the main report
            For Each oTable In oCrystalReport.Database.Tables
                With oTable.LogOnInfo.ConnectionInfo
                    .ServerName = gsDBDataSource
                    .UserID = gsDBUserID
                    .Password = gsDBPassword
                    For Each oPair In .LogonProperties
                        If UCase(CStr(oPair.Name)) = "DATA SOURCE" Then
                            oPair.Value = gsDBDataSource
                            Exit For
                        End If
                    Next oPair
                End With
                oTable.ApplyLogOnInfo(oTable.LogOnInfo)
            Next oTable
            '-- Set database connection info for each table in each subreport
            For Each oSubreport In oCrystalReport.Subreports
                For Each oTable In oSubreport.Database.Tables
                    With oTable.LogOnInfo.ConnectionInfo
                        .ServerName = gsDBDataSource
                        .UserID = gsDBUserID
                        .Password = gsDBPassword
                        For Each oPair In .LogonProperties
                            If UCase(CStr(oPair.Name)) = "DATA SOURCE" Then
                                oPair.Value = gsDBDataSource
                                Exit For
                            End If
                        Next oPair
                    End With
                    oTable.ApplyLogOnInfo(oTable.LogOnInfo)
                Next oTable
            Next oSubreport
    ... alas, it makes no difference.  If we run the report against a database that is different than the one specified with "Set Datasource Location" in Crystal, it fails with nonsense errors 

    Thanks for the reply, Ludek.  We have made some breakthroughs, uncovered some Crystal bugs and workarounds, and we're probably 90% there I hope.
    For your first point, unfortunately the information on the Oracle 6550 error was generic, and not much help in our case.  And for your second point, the errors didn't have anything to do with subreports at that time -- the error would manifest itself even in a simple, one-level report.
    However, your third point (pointing us to KB 1553921) helped move us forward quite a bit more.  For the benefit of all, here is a link to that KB article:
    Link: [KB 1553921|http://www.sdn.sap.com/irj/servlet/prt/portal/prtroot/com.sap.km.cm.docs/oss_notes_boj/sdn_oss_boj_bi/sap(bD1lbiZjPTAwMQ==)/bc/bsp/spn/scn_bosap/notes%7B6163636573733d36393736354636443646363436353344333933393338323636393736354637333631373036453646373436353733354636453735364436323635373233443330333033303331333533353333333933323331%7D.do]
    We downloaded the tool referenced there, and pointed it at a couple of our reports.  The bottom line is that the code it generated uses a completely new area of the Crystal Reports .NET API which we had not used before -- in the CrystalDecisions.ReportAppServer namespace.  Using code based on what that RasConnectionInfo tool generated, we were able gain greater visibility into some of the objects in the API and to uncover what I think qualifies as a genuine bug in Crystal Reports.
    The CrystalDecisions.ReportAppServer.DataDefModel.ISCRTable class exposes a property called QualifiedName, something that isn't exposed by the more commonly-used CrystalDecisions.CrystalReports.Engine.Table class.  When changing the data source with our old code referenced above (CrystalDecisions.Shared.ConnectionInfo.SetConnection), I saw that Crystal would actually change the Table.QualifiedName from something like "SCHEMAOWNER.PACKAGENAME.PROCNAME" to just "PROCNAME" (essentially stripping off the schema and package name).  Bad, Crystal...  VERY BAD!  IMHO, Crystal potentially deserves to be swatted on the a** with the proverbial rolled-up newspaper.
    I believe this explains why we were also able to generate errors indicating that field names or tables were not found -- because Crystal had gone and changed the QualifiedName to remove some key info identifying the database object!  So, knowing this and using the code generated by the RasConnectionInfo tool, we were able to work around this bug with code that worked for most of our reports ("most" is the key word here -- more on that in a bit).
    So, first of all, I'll post our new code.  Here is the main area where we loop through all of the tables in the report and subreports:
    '-- Replace each table in the main report with new connection info
    For Each oTable In oCrystalReport.ReportClientDocument.DatabaseController.Database.Tables
        oNewTable = oTable.Clone()
        oNewTable.ConnectionInfo = GetNewConnectionInfo(oTable)
        oCrystalReport.ReportClientDocument.DatabaseController.SetTableLocation(oTable, oNewTable)
    Next oTable
    '-- Replace each table in any subreports with new connection info
    For iLoop = 0 To oCrystalReport.Subreports.Count - 1
        sSubreportName = oCrystalReport.Subreports(iLoop).Name
        For Each oTable In oCrystalReport.ReportClientDocument.SubreportController.GetSubreportDatabase(sSubreportName).Tables
            oNewTable = oTable.Clone()
            oNewTable.ConnectionInfo = GetNewConnectionInfo(oTable)
            oCrystalReport.ReportClientDocument.SubreportController.SetTableLocation(sSubreportName, oTable, oNewTable)
        Next oTable
    Next iLoop
    '-- Call VerifyDatabase() to ensure that the tables update properly
    oCrystalReport.VerifyDatabase()
    (Thanks to Colin Stynes for his post in the following thread, which describes how to handle the subreports):
    Setting subreport connection info at runtime
    There seems to be a limitation on the number of characters in a post on this forum (before all formatting gets lost), so please see my next post for the rest....

  • Calling two different data sources within same transaction

    Hi,
    I have a method (in an Entity Bean), which I need to use two data sources.
    But the system give this error when I try to do it.
    An illegal attempt to use multiple resources that have only one-phase capability has occurred within a global transaction.
    Please help me to solve this problem.
    Best Regards,
    Chamal.

    I hope you need to use XA Drivers (while create the connection pool it self you need to choose the XA Drivers) for this. else you need to use the userTranscations instead of containers.
    Because other database drivers won't support this feature.
    Please try any of this. Using XA drivers is costly interms of performace.

  • Assigning two different data sources to infosource

    hi
    i have a small problem. Actually we have a development bw server which is having client no.500.
    we have a development r/3 server which is of client 330. and we have defined a rfc connection for client 330 in bw client 500. and all the infosources are assigned to that rfc connection which is configured for 330.
    In 330, we  have very less data. We have one more client 700 which is having data. Now i have created a rfc connection for this client 700. and i want to use this connection  as a source for client 500 in bw.
    can i disable the existing rfc connection of 330 and use rfc connection of 700 for the same infosource.
    Please reply.
    regards
    padma

    You can assign from different clients of your source system datasources to your infosource. I've already linked 30 datasources out of 30 clients to one infosource. No problem.
    So you can link both of your clients to your infosource. I think it is important not to delete the connection to client 330 because this - I think - is the client where you do all the "customizing" setting for your datasource. Then this datasource-infosource connection is the basic for the transports to the production system.

  • Not able to edit the report created on different data source.....

    I have a query regarding Report in OBIEE - reports developed from BI Publisher are specific to data source on which they have been created ??
    i have a sample report that was created on different data source, i have the corresponding RPD also. I changed the data source according to my DB and when i try to update/edit the report,
    on Analytics for adding a new column, it is generating a seperate new Query from QueryBuilder for that additional cloumn rather than adding up the new query with the previous one(existing report query). Is it because of mismatch of data source on which report had been created and on which it is being update ?? if it is the case, where do i need to make changes related to JDBC connection or others ??
    when i try to create a new sample data set and try to update it, it adds up the extra edited things to original query and works perfectly fine. can ny 1 help me for the same ??

    Hi Denis,
    Normally,what we do is once we provide access to webi users group for each user from BO supervisor module, user(s) will able to refresh/edit the existing report from Full client BO.His colleagues have had no problem editing all his reports from their machines but he is not able to edit any report from his machine and BO is getting freeze.
    He also reinstalled BO and cleaned out everything on his Computer and re-built it but the issue is not yet resolved
    Can you please tell me how to resolve this issue
    Kind Regards,
    Srinivas

  • Implementing Logical FACT & Logical dimension from 2 different data sources

    Hi Gurus,
    Here is my situation. We have 2 Different Data sources. One is SRMW and the other one is a different source. What we are trying to implement here is we wanted to create logical dimensions and logical facts as well in the BMM layer. For ex: w_day_d from SRMW and other time dimension from another source makes a logical table. And similarly a fact from SRMW and and another similar fact (Same data types though) will make a logical fact in the BMM.
    I have done the POC of it, but the only problem is that i was able to fetch the data from only one data source not the other source.
    Any suggestions ??
    Thanks in Advance.

    What I already mentioned is that you have to create multiple logical table sources and set the fragmentation content on each logical table source.
    When you have two physical tables for the product dimension, for example DIM_PRODUCT_A and DIM_PRODUCT_B, you must add them to your logical table Products as two separate logical table sources and map all columns to the corresponding logical table columns on the Column Mapping tab.
    Hereafter you should go to the Content tab of each logical table source and describe what content is in the logical table source.
    For example for the logical table source of DIM_PRODUCT_A:
    "BM"."Product"."Product Name" <= 'Product 2'
    and for the logical table source of DIM_PRODUCT_B.
    "BM"."Product"."Product Name" >= 'Product 3'
    Then you must also check "This source should be combined with other sources at this level"
    When you run a query in Answers only on the product table, two queries will be generated to get values from both tables.
    Regards,
    Stijn

  • Different data-sources.xml files for each build

    Any thoughts on how to do this:
    I have two sets of data-sources. When I deploy the EAR to my stage box, I need to go to the "stage" data source of the other repositories. When I deploy the EAR to my production box, I need to use the production data sources for the other repositories
    I am using jdev 10.1.3
    There is a single data-sources.xml file - where, currently, I am manually editing the connection strings between deployments.
    I am using ant to build the EAR file.
    What is a more elegant way to do this?

    Use ANT script to build the EAR file.
    Have 2 different data-sources file one for production and one for stage.
    Based on the environment you want to deploy tell ANT to copy those files into the EAR.
    Just a suggestion.

  • How to process reports with two different date fields

    Morning all,
    This question is somewhat similar to what I asked few days back. I created two different reports for two different departments (Credit Control and Free of Charge Control).
    These two reports pull data from two different date fields and has different status check ups.
    Few days back my director requested me if he can have both reports in one report. That is when I asked question from all of you regarding two reports.
    Anyhow, that was created by implementing two sub reports however, the director asked me to combine the two different reports into one in such a way that the output records should be added up with each other hence one single report with all combined data of Credit Control and Free of Charges (not two sub reports in one report).
    The problem here which I am getting is that, both reports uses two different date fields. I can live with two different statuses as that can be easily done within formula/RT however how to create a report with Report Selection Formula condition which looks in both date fields and produces results accordingly?
    I have tried this by not putting any condition in report selection formula and ran the report, it produced total results from past 6 years.
    Any ideas how to produce combined report which has two different date fields?
    Many thanks in advance
    Regards
    Jehanzeb
    What I

    Charliy,
    No the both dates are from two different databases, tables and fields. The Credit date field is from KLAB database under max table where as the FOC date field is from SAE database under maxmast table. Klab date is under credit2 table where as foc date is under order header table.
    So its more like this:
    Klab->Max-Credit2--->Credit_date
    Sae->Maxmast->Order_header---Date entered.
    two dates totally different tables, dbs
    however, I will try your idea too and see if that works.
    Regards
    Jehanzeb

  • Need suggestion for ISE distributed deployment model in two different data centers along with public certificate for HTTPS

    Hi Experts,
    I am bit confused about ISE distributed deployment model .
    I have two data centers one is DC & other one is as a DR I have  requirement of guest access service implementation using CWA and get public certificate for HTTPS to avoid certificate error on client devices :
    how do i deploy ISE persona for HA in this two data centers
    After reading cisco doc , understood that we can have two PAN ( Primary in DC  & Secondary in DR ) like wise for MnT (Monitoring will be as same as PAN ) however I can have 5 PSN running in secondary i.e. in DR ISE however I have confusion about HA for PSN .. since we have all PSN in secondary , it would not work for HA if it fails
    Can anybody suggest me the best deployment solution for this scenario ?
    Another doubt about public certificate :
     Public Certificate: The ISE domain must be a registered or part of a registered domain name on the Internet. for that I need Domain name being used from customer .
    Please do correct me if I am wrong about certificate understanding :
    since Guest will be the outside users , we can not use certificate from internal CA , we need to get the certificate from service provider and install the same in both the ISE servers
    Can anybody explain the procedure to opt the public certificate for HTTPS from service provider ? And how do i install it in both the ISE servers ?

    Hi there. Let me try answering your questions:
    PSN HA: The PSNs are not configured as "primary" or "secondary" inside your ISE deployment. They are just PSN nodes as far as ISE is concerned. Instead, inside your NADs (In your case WLCs) you can specify which PSN is primary, which one is secondary, etc. You can accomplish this by:
    1. Defining all PSN nodes as AAA radius servers inside the WLC
    2. Then under the SSID > AAA Servers Tab, you can list the AAA servers in the order that you prefer. As a result, the WLC will always use the first server listed until that server fails/gets reloaded, etc. 
    3. As a result, you can have one WLC or SSID prefer PSN server A (located in primary DC) while a second WLC or SSID prefer PSN server B (located in backup DC)
    Last but not the least, you could also place PSNs behind a load balancer and that way the traffic would be equally distributed between multiple PSNs. However, the PSN nodes must be Layer 2 adjacent, which is probably not the case if they are located in two different Data Centers
    Certificates: Yes, you would want to get a public certificate to service the guest portal. Getting a public/well known certificate would ensure that most devices out there would trust the CA that signed your ISE certificate. For instance, VeriSign, GoDaddy, Entrust are some of the ones out there that would work just fine. On the other hand, if you use a certificate that was signed by your internal CA, then things would be fine for your internal endpoints that trust your internal CA but for any outsiders (Guests, contractors, etc) that do not trust and do not know who your internal CA is would get a certificate error when being redirected to the ISE guest portal. This in general is only a "cosmetic" issue and if the users click "continue" and add your CA as a trusted authority, the guest page would load and the session would work. However, most users out there would not feel safe to proceed and you will most likely get a lot of calls to your helpdesk :)
    I hope this helps!
    Thank you for rating helpful posts!

  • How to connect to different data sources in  XI 3.1 UNIVERSE?

    hi experts,
    We are going to migrate from BO 6.5 to  BO XI3.1.
    and we have different data sources to extract and generate the reports using various tools like Webi,Crystal Reports, Deski etc.
    1         How to connect to different data sources like SAP BI query, DB2, Oracle etc,, provide me the step by step solution.
    2     We need to migrate the universes and reports from BO 6.5 to BO XI 3.1 using the Report Conversion Tool.
    can we directly convert from the deski reports to Webi or do we need to migrate first through Import Wizard?
    3    Scheduling the Reports through Infoview
    4    Various issues and how to resolve those issues, we generally face while scheduling, and while creating the queries like for example time out errors etc..
    5    BO Security
    6    Universe creation with best practices and the loops,contexts,aliases etc..
    please provide me the step by step solution documents and links.
    thans in advance
    venuscm

    I would recommend to take a look at the official documentation. YOu can find this here
    http://help.sap.com
    NAvugate to SAP BusinessObjects->All products and the select the product you want eg. BusinessObjects ENterprise or Universe designer and the appropriate version XI 3.1. Take a look at the admin and the user guides.
    Currently the Data FEderator will allow you to buid a universe that access data from various data sources. Another option is to merge queries from different universes at report level.
    Regards,
    Stratos

  • How do I calculate days from two different dates?

    Hi all,
    How do I calculate days from two different dates?
    my requirement is to pass the number of days to target field from two dates.
    Current date :  14/04/2010
    Standard date: 01/01/1957 is the standard = day 0
    Is it possible in graphical mapping with out  udf?
    Plz help me on this I have donu2019t have much knowledge on JAVA.
    Thanks

    Hi Yadav,
    Probably this is not the correct forum for XI / PI .
    You can post the same to...
    Process Integration (PI) & SOA Middleware
    Hope this helps.
    Regards
    Raj

  • How do I calculate days from two different dates in XI/PI?

    Hi all,
    How do I calculate days from two different dates in XI/PI?
    my requirement is to pass the number of days to target field from two dates.
    Current date : 14/04/2010
    Standard date: 01/01/1957 is the standard = day 0
    Is it possible in graphical mapping with out udf?
    Plz help me on this I have donu2019t have much knowledge on JAVA.
    Thanks

    Hi Yadav,
    Probably this is not the correct forum for XI / PI .
    You can post the same to...
    Process Integration (PI) & SOA Middleware
    Hope this helps.
    Regards
    Raj

  • How to calculate AGE from two different date fields

    hi
    I need to calculate AGE from two different date fields.
    Can some help me how to do, when i try to do substraction formula it is showing error.
    Thank You
    Manu

    Manu wrote:
    Hi
    Both fields are in date format only, i need to calculate no . of days between two different dates
    Thanks
    ManuThe reason for the question about the format of the column is because the simplest solution only works on DATE columns. Others have mentioned this here, but if you want the difference between two date fields, you can use this formula:
    TIMESTAMPDIFF(SQL_TSI_DAY, date_column1, date_column2)
    The above being said, you still didn't tell me what you did, or what error message you received. So again, if the above formula didn't work, what did you do? Where did you put the formula? What was the exact syntax you used? What was the error message you got?
    Please don't make us work more than we need to. Answer all the questions in your next post. Thanks.

Maybe you are looking for