Question about 'Generic Database Connector'

Hi SAP IdM Gurus,
I found 'Generic Database Connector' in the list of SAP NW IdM connectors overview.
I'd like to check the capability of this connector in detail. 
I'm wondering whether it could provide not only  a connectivity of RDBMS, but also custom-developed application using the RDBSM for their ID/Password repository. If it provides only a connectivity of RDBMS, Is there a connector for the custom-developed application using the RDBSM for their ID/Password repository?
My question is that which connector is used for the connectivity  of custom-developed application using the RDBSM for their ID/Password repository.
I would really appreciate it if you could answer the question.
Thanks
Kenneth

Kenneth,
The Generic Database connector uses a JDBC connection and allows for the and writing to databases.
That being said, you can also use this connection to set up access to the system or drop access using SQL commands and set them up per the framework.
What specific Database are you working with?
Matt

Similar Messages

  • Question about import database pk, fk

    Hi
    I want to ask a question about imported database. I imported database and I noticed tables doesnt have Primary key or foreign key. They havent relation between eachs. Why can this occured ?
    is it about import or another something?

    yeah ı am surprised none of tables have pk or fk ..
    but ı want to show you http://i51.tinypic.com/2hedc9d.jpg
    what is blue shapes in there ?
    can be foerign keys ?

  • Basic question about Flashback Database

    Hi,
    I have a very generic question about using Flashback Database.
    On my testing systems, for performance testing and simulation purposes, I want to create a guaranteed restore point so I can test impact on batch when code change releases are done, before deployment in production.  My confusion is with respect to redo logs, as summarized in the questions below:
    1. Is it possible to change redo log files, when a guaranteed restore point has been configured?
    2. If yes, will the Flashback to restore point, also change the size of the redo logs?
    I could not find anything in the docs about this....hence my question....
    Appreciate your time taken in responding to these questions....
    Regards.

    HI,
    donneskold wrote:
    Hi,
    I have a very generic question about using Flashback Database.
    On my testing systems, for performance testing and simulation purposes, I want to create a guaranteed restore point so I can test impact on batch when code change releases are done, before deployment in production.  My confusion is with respect to redo logs, as summarized in the questions below:
    1. Is it possible to change redo log files, when a guaranteed restore point has been configured?
    1) Yes it is possible.
    2. If yes, will the Flashback to restore point, also change the size of the redo logs?
    2) It will not change the size of redo log file.  Cannot resize a redo log file..
    I could not find anything in the docs about this....hence my question....
    Appreciate your time taken in responding to these questions....
    Regards.
    Thank you

  • Few questions about upgrading database

    Hi everyone,
    greetings of the day
    I have few questions about the upgrading database,
    In export and import mode
    1.can we have new name for the target database,
    2.I think we need to create tablespaces ,do we need to create users as well
    3.If we are upgrading from 9i to 10g database ,any activity to be perfromed other than creating a new sysaux tablespace
    4.How come we get consistent export
    In DBUA mode ( in the same machine only)
    1.Do we need shutdown / startup restrict the database
    2.How can we move the files to the new location
    3.Can we change db_name of the database
    4.Can we use the old database as well
    In manual upgration using catupgrd scripts
    1.Can we rename the db_name
    2.can we use old database as well
    3.how to move the database files to the new location
    4.can we perform this kind of upgrade on a different server
    5.when we startup upgrade in the new home ,how it identifies the old database in-order to upgrade the old one
    Thanks

    udayjampani wrote:
    Hi everyone,
    greetings of the day
    Pl post details of source and target database versions, along with your OS details.
    I have few questions about the upgrading database,
    In export and import mode
    1.can we have new name for the target database,Yes.
    2.I think we need to create tablespaces ,do we need to create users as wellYou can create users, but it is not necessary. You need to pre-create tablespaces only if their characteristics/locations on the target are different than on the source.
    3.If we are upgrading from 9i to 10g database ,any activity to be perfromed other than thisNot that I am aware of - see the steps in the Upgrade Guide - http://docs.oracle.com/cd/B19306_01/server.102/b14238/expimp.htm
    4.How come we get consistent export wetherEnsure the database is started in restricted mode, so users will not be able to access the database during the export.
    >
    In DBUA mode ( in the same machine only)
    1.Do we need shutdown / startup restrict the databaseNo - DBUA will do this automatically for you.
    2.How can we move the files to the new locationYou can after the upgrade move the datafiles to wherever you want - use the ALTER DATABASE RENAME DATAFILE (http://docs.oracle.com/cd/B19306_01/server.102/b14200/statements_1004.htm#i2082829) command.
    3.Can we change db_name of the databaseI do not believe this is possible with DBUA.
    4.Can we use the old database as wellNo - the database will be upgraded by DBUA - there is no "old" database.
    >
    >
    In manual upgration using catupgrd scripts
    1.Can we rename the db_nameYes.
    2.can we use old database as wellNo - the scripts will upgrade the database - there is no "old" database.
    3.how to move the database files to the new locationSee above.
    4.can we perform this kind of upgrade on a different serverPl elaborate on what you mean by this. You can copy the existing database to a different server (assuming compatible OS) and upgrade it there.
    >
    >
    ThanksHTH
    Srini

  • Basic questions about Notes domino connector

    Hi,
    I have a very basic question about the Lotus Domino connector for OIM. We have a requirement to provision accounts into the domino mail server and while doing that also update some information directly into a separate nsf file.
    I checked the documentation and it seems while configuring an IT resource, we need to give the Mail DB path, and thats the only place where it refers to it.
    Could you guys please let me know if this can be done? Can I write directly to a different nsf file, which is not like provisioning to the complete mail server. So the other attributes in the IT resource may not be able to be set up as its just the file I need to provision into.
    Looking forward to your response.
    Thanks, M

    Hello folks, please share your experience working with the lotus notes connector.
    Thanks, M

  • A Question about LV Database Connectivi​ty Toolkit

    Hello everyone!
    I have a question about using LabVIEW DataBase Connectivity Toolkit 1.0.2 that eagerly needs your help. I don't know how to programmaticlly create a new Microsoft Access(.mdb)file (Not a new table in a existing Database)using LabVIEW Database Connectivity Toolkit1.0.2. As you know, usually we can set up the connection by creating a Universal Data Link (.udl) file and inputting the path to the DB Tools Open Connec VI in the LabVIEW DataBase Connectivity Toolkit. However, searching a table within an existing database containing a great many tables is a toilfulif job. If I want to use a new DataBase file with the date and time string as its name to log my acquisition data in each measurement process, how to do? I am sure someone of you must can resolve my question, and thanks very much for your help.

    I don't know what your real design considerations are here but, from I understand from your post, this is a really bad way to go about the process of logging data -- IF you want to be able to do significant ad hoc or stored procedures analyses after it has been collected.  Using separate MDB files for data that ONLY differs by one field (namely that date) is not the most efficient way to organize it.  What would be much more efficient would a joined table including the date and a reference ID of some sort for the various measurements that were done.  That way your stored procedures for looking at ALL measurements of type X would be very simple going across ALL dates.  Making such a comparison across multiple MDB files is a much more challenging process AND doing the original data collection in that way doesn't really gain you anything.
    Generally, if something is difficult to do in the DCT (Database Connectivity Toolkit) it's because it's a "not good thing" to do within MDBs.  I know that others probably disagree with that but I've worked with Access since it's initial release and other RDBMs prior to that both through compiled tools, Unix scripts, etc.  You may, of course, still choose to proceed in the way you've described and that may work excellently for you. 

  • Question about Oracle Database 11g: Program with PL/SQL

    Hello,
    I've read that:
    This course is a combination of Oracle Database 11g: PL/SQL Fundamentals and Oracle Database 11g: Develop PL/SQL Program Units courses.
    In order to preparing for this exam, i should read the student guide, so there are 4 pdf files. One is Oracle Database 11g: PL/SQL Fundamentals, and the other 3 are regarding to Oracle Database 11g: Develop PL/SQL Program Units (volume I, II and III). My question is, regarding to this last one (Oracle Database 11g: Develop PL/SQL Program Units), which volume should i read? I saw that volume I has 400+ pages, and the IIIrd ~80.
    Which document should i read? What's the difference between them?
    Thanks!

    Passing the exam and becoming certified are not a matter of simply reading a book or taking a training class. While books and classroom materials as well as in-class training are all important steps in preparing for an exam, experience and research are just as important. I cannot tell you that there are 53 questions on the exam and every single question maps directly back to page 13, paragraph 3 in XX book. However, If you have the classroom material, and you have the list of exam topics, you should be able to find most of the information you need. In some cases, if you are not completely familiar with a specific topic, or you don't feel like the information in the book gives you all the information you need, you may need to do a little additional research. This is all part of preparing for an exam. This process is what makes you successful as a certified professional.
    Your focus seems to be solely on the course materials that you have and what volumes and chapters you should read. I'm trying to give you a different perspective - preparing for an exam is not necessarily about taking one course or reading one book and being ready. You need to review the exam topics and be sure that you are familiar with each topic.
    Regards,
    Brandye Barrington

  • Question about B1if-Database in SQL-Mgmt-Studio

    Hello,
    We have made some changes in the configuration in SLD.
    We made some more Di-Proxy-Ports and then made Proxygroups for the Databases.
    Nearly every half hour, in SQL-Mgmt Studio we can see this "task" ?
    During the time, this task is online, the Queue-Monitor collects all Jobs... but not more.
    After this Job has finished... B1i begins to work up the jobs.
    Can anyone tell me something about "Rollback Transac" in  B1if-Database ?
    Thanks
    Markus

    Passing the exam and becoming certified are not a matter of simply reading a book or taking a training class. While books and classroom materials as well as in-class training are all important steps in preparing for an exam, experience and research are just as important. I cannot tell you that there are 53 questions on the exam and every single question maps directly back to page 13, paragraph 3 in XX book. However, If you have the classroom material, and you have the list of exam topics, you should be able to find most of the information you need. In some cases, if you are not completely familiar with a specific topic, or you don't feel like the information in the book gives you all the information you need, you may need to do a little additional research. This is all part of preparing for an exam. This process is what makes you successful as a certified professional.
    Your focus seems to be solely on the course materials that you have and what volumes and chapters you should read. I'm trying to give you a different perspective - preparing for an exam is not necessarily about taking one course or reading one book and being ready. You need to review the exam topics and be sure that you are familiar with each topic.
    Regards,
    Brandye Barrington

  • Question about generics and subclassing

    Hi all,
    This has been bothering me for some time. It might be just me not understanding the whole thing, but stay with me ;):
    Suppose we have three classes Super, Sub1 and Sub2 where Sub1 extends Super and Sub2 extends Super. Suppose further we have a method in another class that accepts (for example) an AbstractList<Super>, because you wanted your method to operate on both types and decide at runtime which of either Sub1 or Sub2 we passed into the method.
    To demonstrate what I mean, look at the following code
    public class Super {
      public methodSuper() {
    public class Sub1 extends Super {
      public methodSub1() {
        // Do something sub1 specific
    public class Sub2 extends Super {
      public methodSub2() {
         // Do something sub2 specific
    public class Operate {
      public methodOperate(AbstractList<Super> list) {
        for (Super element : list) {
           // Impossible to access methods methodSub1() or methodSub2() here, because list only contains elements of Super!
           // The intention is accessing methods of Sub1 or Sub2, depending on whether this was a Sub1 or Sub2 instance (instanceof, typecasting)
    }The following two methods seem impossible:
    Typecasting, because of type erasure by the compiler
    Using the instanceof operator (should be possible by using <? extends Super>, but this did not seem to work.
    My question now is: How to implement passing a generic type such as AbstractList<Super> while still making the methods of Sub1 and Sub2 available at runtime? Did I understand something incorrectly?

    malcolmmc wrote:
    Well a List<Super> can contain elements of any subclass of super and, having obtained them from the list, you could use instanceof and typecast as usual.I agree with you on this one, I tested it and this simply works.
    Of course it would be better to have a method in Super with appropriate implementations in the subclasses rather than use distinct method signatures, instanceof and casting isn't an elegant solution. Alternatively use a visitor pattern.Not always, suppose the two classes have some similarities, but also some different attributes. Some getters and setters would have different names (the ones having the same name should be in the superclass!). You want to be able to operate on one or the other.
    Generics doesn't make much difference here, exception it would be more flexible to declare
    public methodOperate(AbstractList<? extends Super> list) {Which would alow any of AbstractList<Super>, AbstractList<Sub1> or AbstractList<Sub2> to be passed.I tried it and this also works for me, but I am still very, very confused about why the following compiles, and gives the result below:
    public class Main {
         public static void main( String[] args ) {
              AbstractList<Super> testSub = new ArrayList<Super>();
              testSub.add( new Sub1( "sub1a" ) );
              testSub.add( new Sub1( "sub1b" ) );
              accept( testSub );
         private static void accept( AbstractList<? extends Super> list ) {
              for ( int i = 0; i < list.size(); i++ ) {
                   Super s = list.get( i );
                   System.out.println( s.overrideThis() );
    public class Sub1 extends Super {
         private String sub1;
         public Sub1( String argSub1 ) {
              sub1 = argSub1;
         public String overrideThis() {
              return "overrideThis in Sub1";
         public String getSub1() {
              return sub1;
    public class Sub2 extends Super {
         private String sub2;
         public Sub2( String argSub2 ) {
              sub2 = argSub2;
         public String overrideThis() {
              return "overrideThis in Sub2";
         public String getSub2() {
              return sub2;
    public class Super {
         public Super() {
         public String overrideThis() {
              return "OverrideThis in Super";
    }With this output:
    overrideThis in Sub1
    overrideThis in Sub1Even using a cast to Super in Main:
    Super s = (Super) list.get( i );does not return the Sub1 as a Super (why not??)
    Sorry for the long code snippets. I definitely want to understand why this is the case.
    (Unwise, I thing, to use Super as a class name, too easy to get the case wrong).OK, but I used it just as an example.

  • Another question about generics

    Consider the following hierarchy of classes:
    class Fruit {...}
    class Orange extends Fruit {...}
    class Apple extends Fruit{...}
    Now I need to declare a collection wherein I could store any type of Fruit (including its derivated classes), but also gather some fruits.
    I hare read that for reading I need to declare the collection as follows:
    Set<? extends Fruit> fruitSet = new HashSet<Fruit>();
    and for writing I need the following declaration:
    Set<? super Fruit> fruitSet = new HashSet<Fruit>();
    My question is, is there an efficient way to declare the collection so that both operations be feasible in the class? In the current version, I use generic method and need to re-cast for some calls.
    Can anyone help?

    What they're talking about is the fact that parameterized types are not covariant. You can't treat a List<Rectangle> as a List<Shape> because they are not the same thing. Arrays on the other are covariant, you can treat a Rectangle[] as a Shape[].
    No, with your example it will not generate a compile time error, but I do see what you're getting at.
    public void readFruit(Set<Fruit> fSet) {
        // snip
    Set<Apple> fruits = new HashSet<Apple>();
    readFruit(fruits); // error because Set<Apple> is NOT a Set<Fruit>It has nothing to do with whether or not fruits contains Apples and Oranges, it's how fruits is declared. While a Set<Fruit> may contain Apples and Oranges a Set<Orange> is not a Set<Fruit>. So if you attempt to pass a Set<Orange> to a method that takes a Set<Fruit> you might as well be trying to pass an int to a method that accepts a String, they're not the same, there is no is-a relationship. So this is what you do:
    public void readFruits(Set<? extends Fruit> fSet) {
        // snip
    // All work because it will take any Set<E> where E extends Fruit.
    readFruits(new HashSet<Fruit>());
    readFruits(new HashSet<Apple>());
    readFruits(new HashSet<Orange>());What this says is that fSet can be any Set<E> where E is a subtype of Fruit. It doesn't matter what implementations the Set actually contains. If it's a Set<Fruit> and it contains Apples, it doesn't matter, it's simply allowing you to pass a Set<Apple> or Set<Orange> as opposed to only being able to pass a Set<Fruit>.

  • Maybe a dumb question, about which power connectors to use, but....

    This may sound like an incredibly dumb question, but here goes anyway....
    I upgraded my system a few weeks ago, and will be slapping in a new ATI X850 Pro AGP card in a few days.  My current graphics card is a p.o.s. Radeon 9550, so it doesn't have its own power connector or even a fan.  My PSU is a dual-rail Antec model, and I'm wondering if it should matter any about how many or which devices I have connected to each power cable set.  My psu came with cables to attach, only 2 of the 4 though are the 4-pin molex - the other 2 are pci-e power connectors I believe.  Should it make a difference which or how many devices I connect to each cable?  For ease of plugging in, I believe my DVDRW and CDRom are using 1 cable, and my ide drive and floppy are using the other cable. Also, I have a pci fan and a case fan running off 1 or the other.  I'm likely to jump the power cable for the video card off the one the hard drive is using, unless someone has some suggestion to offer to the contrary.
    If it makes a difference, can anyone recommend how they would have them connect?
    My list of devices to plug in would be:
    1)IDE HDD
    2)Floppy
    3)CDRom
    4)DVDRW
    5)Case Fan
    6)PCI Fan
    7)Video Card
    Thanks for any thoughts!  SAB - you seem to be the man on builds.....any ideas? 

    Well, I put the new card in and connected it on a cable with the optical drives.  Also, I had unwittingly connected my cdrom on the same ide channel as my hdd    , so I changed that so both optical drives are on the secondary ide.
    So far, so good.  3dMark03 jumped up from 3300 with an overclocked Radeon 9550 to 10570 with the stock Radeon 850 Pro 
    I wish I could find some local source to purchase additional molex cords for my Antec SmartPower 2.0 psu, though - I've got 2 more 'outlets' just sitting there on the psu doing nothing, because the other cables are for pci-e, and I'm not using pci-e......  If I could buy some additional molex cables to plug in like the ones that came with it, I could have the vid card on its own cable....

  • Question about using database links in Forms 10g

    I have a form that uses a database link for the "Query Data Source Name," i.e. it is set to some_view@some_db_link
    For some reason when I run a query using the DB link the query takes approximately 30 seconds. When I remove the link and go local the query takes about one second. Has anybody observed and addressed this issue before (without going local of course)?
    Thanks,
    Thomas

    Let me first say that it is never a good idea to use a db link in the data source. It is better to create a synonym in the database and use this as the data source. If the db link changes, you only have to do this in the database, without recompiling forms.
    create synonym some_view for some_view@some_db_link
    This may already solve your problem. If not, than see if the query is also slow in sqlplus. You may have a problem in your connection between the two databases.

  • Some questions about my database access bean

    The following is a bean (minus imports) that access a database and populates several "articles" and then returns a set of them. pretty plain really.
    The bean is modified from here to add exception cactching and 'fix' other problems along the way.
    http://forum.java.sun.com/thread.jspa?threadID=727979 (thanks evnafets)
    The results list is genericified(!) to avoif unchecked calls (casting) in this line result.add(article);
    public class dataAccess {
        public List getNewsResults() throws NamingException{
            Context initContext = new InitialContext();
            Context envContext = (Context) initContext.lookup("java:/comp/env");
            DataSource ds = (DataSource) envContext.lookup("jdbc/mySite");
            Connection conn = null;
            Statement stmt = null;
            ResultSet rs = null;
            List<newsArticleBean> result = new ArrayList<newsArticleBean>();
            try {
                String sql = "SELECT * FROM news, users WHERE news.user = users.username";
                conn = ds.getConnection();
                stmt = conn.createStatement();
                rs = stmt.executeQuery(sql);
                stmt.execute();
                while (rs.next()) {
                    newsArticleBean article = new newsArticleBean();
                    article.setTitle(rs.getString("title"));
                    article.setAuthor(rs.getString("author"));
                    article.setEmail(rs.getString("email"));
                    article.setNews(rs.getString("news"));
                    article.setPictureURL(rs.getString("pictureURL"));
                    article.setDate(rs.getDate("date"));
                    article.setTime(rs.getTime("time"));
                    result.add(article);
            } catch (SQLException e) {
                //error message goes here
            finally {
                try {
                    if (rs != null) rs.close();
                } catch (Exception e) {
                    //error message goes here
                try {
                    if (stmt != null) stmt.close();
                } catch (Exception e) {
                    //error message goes here
                try {
                    if (conn != null) conn.close();
                } catch (Exception e) {
                    //error message goes here
        return result;
    }Some questions that would be of great help if answered:
    Im not too sure what a context does or what im doing with it here. My current understanding is that it allows concurrent access of the database.
            Context initContext = new InitialContext();
            Context envContext = (Context) initContext.lookup("java:/comp/env");
            DataSource ds = (DataSource) envContext.lookup("jdbc/mySite");This section im having trouble with too.
    As I see it:
    I set what my sql query will be;
    get a connection @ conn
    now im not too sure what stmt is getting, is it getting ready to send the query.
    stmt does not have an executeQuery() method, so im never actually sure when the sql query gets sent to the database. I have subbed with execute() but it does not take any arguments so the problem of not sending anything remains.
                String sql = "SELECT * FROM news, users WHERE news.user = users.username";
                conn = ds.getConnection();
                stmt = conn.createStatement();
                rs = stmt.executeQuery(sql);Any help greatly appreciated.

    Some questions that would be of great help if
    answered:
    Im not too sure what a context does or what im doing
    with it here. My current understanding is that it
    allows concurrent access of the database.
    Context initContext = new InitialContext();
    Context envContext = (Context)
    Context) initContext.lookup("java:/comp/env");
    DataSource ds = (DataSource)
    aSource) envContext.lookup("jdbc/mySite");
    Nops, it has nothing to do with concurrent access. You use tomcat to set up a Connection Pool. This causes the tomcat server to create a pool of reusable connections on startup. The connection pool (or set of Connection objects) are given a name and the server binds the set of objects to this name in a 'registry'.
    From the code, you use jndi to access the Connections from the pooled objects. While doing so, you have to provide the jndi name against which this object pool is bound to.
    You'll be more clear when you read the fllowing resources.
    http://java.sun.com/products/jndi/
    http://tomcat.apache.org/tomcat-5.5-doc/jndi-datasource-examples-howto.html
    >
    This section im having trouble with too.
    As I see it:
    I set what my sql query will be;
    get a connection @ conn
    now im not too sure what stmt is getting, is it
    getting ready to send the query.Yes something of that sort - basically it provides you a wrapper which can accept your query, use the connection object to access the db, execute the query and return a result set.
    stmt does not have an executeQuery() method, so im
    never actually sure when the sql query gets sent to
    the database.er....it has and that's what you use in your code. (http://java.sun.com/j2se/1.5.0/docs/api/java/sql/Statement.html#executeQuery(java.lang.String))
    I have subbed with execute() but it
    does not take any arguments so the problem of not
    sending anything remains.Hmmm....the execute() method is for a different purpose and it does accept arguments (http://java.sun.com/j2se/1.5.0/docs/api/java/sql/Statement.html#execute(java.lang.String))
    >
                String sql = "SELECT * FROM news,
    users WHERE news.user = users.username";
    conn = ds.getConnection();
    stmt = conn.createStatement();
    rs = stmt.executeQuery(sql);Any help greatly appreciated.ram.

  • Question about a database saving

    Hey all. First off, I just started playing with databases yesterday, so forgive my question if it seems obvious/stupid.
    I finally got JDBC working with HSQLDB (I wanted something portable), and wrote some code to connect to the database, create a table, and input/retrieve data from the table. The only way I could get the database to retain the stuff I wrote if I opened it the next time was to use statement.execute("SHUTDOWN"). Now this is fine, but is there any way to have the database save all the info besides using the shutdown command?
    When I didn't use the shutdown command, the next time I connected to the database and tried to read values from the table I made before, I got an error. When I did use the shutdown command, it printed them right out for me, so I'm assuming that IS my problem (It very well may not be).
    Anyway, I had a hard time explaining, so if any further explaination is needed please let me know. And thank you in advance!
    Edit: If sample code would be nice, just say so and I will post a couple examples of the issue.

    Alright. I looked through both. The commit thing will come in very handy, but I think I described the problem in slightly the wrong way. It is my understanding that autoCommit is set to a default of true. So I will set it to false and try to make my program employ ACID (Which might be kind of hard, I'm new at this). But I believe my testing program had a slightly different problem. Since there's not much of a way to distribute my current database, I'll try to describe the problem instead of posting the code.
    I have a databse with a table created. It can hold an employee name as well as their ID number (Not for my real program, I was following a tutorial). I had one simple add statement, that added an employee name and an ID. If I run the program with
    st.execute("SHUTDOWN");at the end, every time I run the program I will get an additional copy of that data every time I run it. Without the shutdown, I can run the program 100 times and still only have as many entries as I had before.
    That's kind of what I meant by the database not saving. It may be commiting it, but if I close without the shutdown command, it doesnt seem to change the database.
    Am I coding something incorrectly, or is this intended behavior? The question may seem slightly weird and useless, but what if someone alt-f4ed out of a program? I don't see how the shutdown command could be issued then. Again, sorry if my question is hard to understand, describing database problems isn't easy for me.

  • A question about Generic Extractor

    Hi everyone,
    I excuted Generic Extractor to extract customer reference data to MDM server. I deselect "Local download" and  select "Upload via FTP", there is no error and the reference data was transfered to corresponsing mdm port directories. But xml files of the reference data were also downloaded into the local directory from where I upload Ports and Check-Tables.
    Could you give some suggestion about it.
    best regards,

    Hi Lionsir,
    It could be the case that you have given the local file directory address in the generate XSD option.
    Just check the same.
    Hope it helps.
    Thanks,,
    Minaz

Maybe you are looking for

  • "External Data Refresh Failed" in Excel Workbook in SharePoint 2013

      Hi Everyone, I hope someone can help me sort this out. I wasn't sure if this should be submitted to SQL Server 2012 or SharePoint 2013 but I think I have ruled out an actual SQL issue here. I have a test SharePoint 2013 Farm installed on one server

  • PO an dPO line item details in FBL3n or FAGLL03 display

    Dear All, I have an requirement where in the user is expecting the clearing of GRIR items , for clearing we are using the standard transction code regularly - but what happens every time there will be some uncleared entries due to some reason - now t

  • IDVD chapter headings anamorphic

    According to Apple's technical note T51611, an anamorphic 16:9 sequence must be exported from Final Cut Express to iDVD using certain custom settings from the File>Export>Using Quicktime Conversion menu. However, this bypasses the "Include Chapter He

  • AMD Opteron on Oracle 9i.

    Hi, I would like to know if AMD Opteron platfrom is now supported on Oracle 9i using SUSE or Redhat Linux distribution. If not, do we know when will be it available or any plans for this platform? Any info would much be appreciated. marlonie

  • Where can I get a copy of Photoshop cs 5.1?

    My old Win computer crashed and I need to reinstall.  I tried the link http://prodesigntools.com/all-adobe-cs5-direct-download-links.html for the download but got a permissions error Access denied, even though logged in.  I've got the Serial number t