Understanding Datasources

Hi gurus,
I have an issue in seecting the DS in DS expert, I am refering to a report in building a new report with slight modifications on same DB.
Firstly, I am in satandard report creation wizard - what is the diffrence of Datasources's - Dataroot/Variance_Xtab and Variance_Xtab (XML Webservices) (probably a naming convetion query)
And; I assumed both are same and ran the report with formatting with some sample objects and at execution it gave me some formula error as the data has to be same if DS's are same.
Please assit me in understanding the DS basing on the tech name and formula issues diagnising
Thanks in advance,
Bhanu

No idea what you are asking or talking about so I assume you are using Classic SAP data sources?
Is this Crystal Reports Designer you are using?
What is a DS in DS Expert?
I suggest you post your question to the SAP Integration Kit forum or possible the B1 forum.
Don

Similar Messages

  • Help me understand JDBC connections using DataSource objects

    I'm writing a simple Java server application that accepts connections from multiple clients via RMI, connects to a SQL Server 2000 database via Microsoft's SQL 2000 JDBC driver, gets some data, and returns it to the client. The server application will handle every SQL database connection. However, it must be able to establish multiple simultaneous connections, since multiple clients may connect at the same time and request data.
    Sounds like a simple multi-threaded server that utilizes a connection pool for database connectivity, right? Well, if I want to do that, I have to register a PooledDataSource object with JNDI. My server threads that are handling requests will simply grab connections out of the pool to retrieve data (presumably).
    This seems like a fairly simple server application; however, all the information and tutorials I've read on using DataSource objects for database connections rather than DriverManager.getConnection() mention XML config files, J2EE application servers, Tomcat, WebSphere, etc. I really don't want to have to deal with all that just to get connection pooling and not have to use the DriverManager.getConnection() method.
    Is there a simple way to use DataSource objects without installing a lot of extra junk on my server?

    Thanks, diazlara. Your response was exactly what I needed to know. All the information I've read prior to this suggested that I had to have a J2EE server running, and that I had to register the DataSource object with JNDI. If you ladies and gents could take a quick look at this and let me know if I'm doing it correctly, I'd appreciate it.
    As proflux suggested, I wrote a connection manager:
    import java.sql.*;
    import javax.sql.*;
    import com.microsoft.jdbcx.sqlserver.*;
    public class SQLConnectionManager {
       private static ConnectionPoolDataSource connectionPool;
       // Initialize the connection manager
       public SQLConnectionManager() {
          init();
       // Retrieve a connection from the connection pool and return it
       public static Connection getConnection() {
          Connection connection = null;
          try {
             connection = connectionPool.getPooledConnection("myUsername", "myPassword").getConnection();
          catch (SQLException ex) {
             System.err.println("A problem occurred while getting a connection from the pool.");
             ex.printStackTrace();
          return connection;
       // Initialize the datasource and connection pool
       private static void init() {
          SQLServerDataSource mds = new SQLServerDataSource();
          mds.setDescription("MS SQLServerDataSource");
          mds.setServerName("127.0.0.1");
          mds.setPortNumber(1433);
          mds.setDatabaseName("SomeTable");
          mds.setSelectMethod("cursor");
          connectionPool = (ConnectionPoolDataSource) mds;
    }As you can see, I'm hard-coding the username and password for testing purposes. I'll incorporate a login/authentication system for any production code that I use this in.
    Once the SQLConnectionManager is created, connections can be retrieved as follows:
          Connection connection = null;
          try {
             connection = SQLConnectionManager.getConnection();
             System.out.println("** Got a SQL database connection from the pool");
             DatabaseMetaData metaData = connection.getMetaData();
             ResultSet rs = metaData.getCatalogs();
             while (rs.next()) {
                System.out.println("\tcatalog: " + rs.getString(1));
             rs.close();
             connection.close();
          catch (SQLException ex) {
             System.out.println("An exception occurred while connecting to the SQL Database:");
             ex.printStackTrace();
          }So, how does that look? I tested it out, and it seems to work just fine. Using this code, I am getting a connection from a pool, correct? I want to make sure of that. Also, are there any inherent weaknesses or flaws with this method that I need to be aware of?
    Thanks so much for all the helpful responses.

  • Creation of DataSource for FlatFile Interface in BI 7

    Hi Y'all!
    Can anyone please give me urgent assistance in creating/defining a DataSource for the FlatFile Scenario in BI 7.0.  I've been trying to no avail.  I certain I am doing this incorrectly, as such I seek your assistance.  Can anyone please be so kind as to give me a step-by-step walkthrough of the specific individual steps (in sequence) in creating/defining DataSources...  Also can this be accepted as a character (08-MAY-07)?
    Much thanks!
    Philips

    Hi,
    The following step can help you..
    1. create a info source and assign the data source -
    2. select data source as PC File ( give any name tech & desc)
    3. create a info source structure by include all individual info objects / har. keyfig
        those are available in your file
    4. you can always take the reference of the info objects from the cube/ODS as to
        make info source staructure easy
    5. map the corresponding info objects in the trasfer structure
    6. save and activate the info source
    7. go to your info provider cube/ODS and right click - select - create update rules
        select the name of the file data source name as given above
    8. map them as defult one-to-one mapping until other wise you have some require
        ment in data manipulation through update rules.
    9. save and activate.
    10. come to the info source - create a info package - select source as .CSV file
         and preview it before you actually load.
    11. as far as date is concern in your falt file..select the entire column of the date
         then goto-format column-select 'customize' then put format as 000000 (six
         zeros ) for 0calmonth.
    12. when you lookat the date at infoprovider it shows like 20070508 but
          when you execute in a query it displays its text as 08-May-2007
    I hope it would help you to understand and get your purpose solve.
    Thanks=points.
    Cheers.

  • Enhancement to 0MAT_PLANT_ATTR datasource

    I need to enhance 0MAT_PLANT_ATTR datasource with additional fields STPRS, PEINH from MBEW table and WAERS (currency for STPRS) from T001 table. Table T001 and MBEW table have no fields in common. So I took an indirect approach to this.I have a program for enhancing a BW DataSource in R/3. Here is the code:
    types: begin of zty_waers,
    bukrs type bukrs,
    waers type waers,
    bwkey type bwkey,
    end of zty_waers.
    data: zit_waers type standard table of zty_waers,
    wa_waers like line of zit_waers.
    DATA : it_temp type standard table of biw_marc_s,
    wa_temp like line of it_temp,
    S_INDEX TYPE SY-TABIX,
    z_index type sy-tabix.
    DATA : BEGIN OF ITAB_MBEW OCCURS 0,
    MATNR TYPE MATNR,
    BWKEY TYPE BWKEY,
    STPRS TYPE STPRS,
    PEINH type PEINH,
    WAERS type WAERS,
    END OF ITAB_MBEW.
    data: wa_mbew like line of itab_mbew.
    DATA: BEGIN OF ITAB_CEPC OCCURS 0,
    PRCTR LIKE CEPC-PRCTR,
    KOKRS LIKE CEPC-KOKRS,
    DATBI LIKE CEPC-DATBI,
    SEGMENT LIKE CEPC-SEGMENT,
    END OF ITAB_CEPC.
    select abukrs awaers b~bwkey into table zit_waers
    from t001 as a inner join t001k as b on
    abukrs = bbukrs.
    IF NOT I_T_DATA[] IS INITIAL.
    IT_TEMP[] = I_T_DATA[].
    SELECT MATNR BWKEY STPRS PEINH FROM MBEW INTO TABLE ITAB_MBEW FOR ALL ENTRIES IN IT_TEMP
    WHERE MATNR = IT_TEMP-MATNR and BWKEY = IT_TEMP-WERKS.
    SELECT PRCTR KOKRS DATBI SEGMENT FROM CEPC INTO TABLE ITAB_CEPC FOR ALL ENTRIES IN IT_TEMP
    WHERE PRCTR = IT_TEMP-PRCTR AND KOKRS = IT_TEMP-KOKRS AND DATBI = '99991231'.
    ELSE.
    EXIT.
    ENDIF.
    loop at itab_mbew into wa_mbew.
    z_index = sy-tabix.
    read table zit_waers into wa_waers with key bwkey = wa_mbew-bwkey.
    if sy-subrc = 0.
    wa_mbew-waers = wa_waers-waers.
    endif.
    modify itab_mbew from wa_mbew transporting waers.
    endloop.
    sort itab_mbew by matnr.
    loop at i_t_data into wa_temp.
    s_index = sy-tabix.
    Read Table itab_mbew with key matnr = wa_temp-matnr
    into wa_mbew.
    if sy-subrc = 0.
    wa_temp-zzstprs = wa_mbew-stprs.
    wa_temp-zzwaers = wa_mbew-waers.
    wa_temp-zzpeinh = wa_mbew-peinh.
    endif.
    Sort itab_cepc.
    Read Table itab_CEPC
    with key PRCTR = wa_temp-PRCTR
    KOKRS = wa_temp-KOKRS
    binary search.
    if sy-subrc = 0.
    wa_temp-ZZSEGMENT = itab_CEPC-SEGMENT.
    endif.
    modify i_t_data from wa_temp.
    clear wa_temp.
    endloop.
    But when I am testing the datasource for the new fields the data is coming out wrong. For example if I run the extractor checker RSA3 for material number 10000306 and plants 1017 - 1021 i extract 2 entries and the stprs values are same for both plants
    But if I run RSA3 for 1021 alone then i get another value for 1021 plant. This means the STPRS value is getting overwritten for some reason. Debugging of this code is very difficult since there is a form embedded in this program.
    Can anyone tell me wht is wrong with the logic?
    Thanks
    Deepthi

    Hi Deepthi,
    As I see and understand your code,
    the problem area could be :
    sort itab_mbew by matnr.  ***
    Here you are sorting the table by MATNR alone, where as when actuall see that
    the remaining data  would be corresponding to the material only, so when sort
    the table alone with MATNR probably your Plant data would get jumble..
    so, try sorting the table with all the keys you are having in this table..
    and read the table with the max. comibations available or possible..
    you should be able to locate the record right get right data..
    Hope it helps..
    Cheers,
    Pattan.

  • DataSource for FAGLFLEXT and BSEG, or New Table in ECC6?

    need to create an extractor to have all the information of FAGLFLEXT, because we need to keep the ledger information and the split of the information. However, we need to add 13 fields contained in BSEG.
    Therefore we thought to reads the line items table FAGLFLEXA, and then enhace it throught BSEG table.
    However, since we are using ECC6 and BI7. It is not support the creation of DataSources for FAGLFLEXA throught FAGLBW03.
    Is it an option to incorporate all fields into FAGLFLEXT.
    Can we creat a new table group based on FAGLFLEXT, and then adding the coding block extensions to that table -
    how does new g/l and the new table group work in parallel? Which is the procedure to do it?
    Documentation says we can create a new table group based on FAGLFLEXT --- its the how does it work in conjuction part...for example...the new g/l handles document splitting and one other thing georg referenced last night...will the split documents go into our new table group?
    BSEG does not have the document splitter information that we need (it's incomplete data). It's missing profit centers on many items, it's missing the proper split of transactions.
    Thanks for your comments.

    Here is more information about this post.
    Client situation:  Our client is implementing ECC 6 and is using the "New-GL" features.  Because of business requirements, the coding block has been extended (not insignificantly - 18 extra fields at the moment) to accommodate legal, regulatory and management reporting.  The reporting solution includes standard ECC reporting (e.g. report writer, report painter reports) as well as feeds to BW (BI 7).
    The Challenge:  Our understanding is that adding all of the coding block extensions to the New-GL tables (ie. FAGLFLEXA and FAGLFLEXT) may lead to performance degradation in the ECC system.  However, we still need to accommodate the requirement to report by the additional dimensions that are not currently included in the New-GL, so our challenge has been to find a solution that minimizes performance issues, while still allowing us to have all the necessary dimensions with which to do the required reporting.
    What we would like to know:  How have you handled this in similar situations?
    Have you added to the New-GL tables? How many fields? Performance issues encountered?
    Have you created additional table group(s) based on the New-GL and then modified that structure to have the new fields?  How does the additional table group work co-incident with New-GL (e.g. does the additional table group receive document splitting information?)?
    Have you created custom extractors for BW?  On what basis (we understand that FAGFLEXA cannot be created as a datasource to feed BW)?

  • Enhance COPA datasource

    I need to enhance a COPA datasource but I cannot add fields in COPA structure as it is already full. So Need to write a code to populate the fields. We require partner function fields from VBPA table and populate the data.
    I cannot create generic delta based on VBPA as this does not contain timestamp field.
    When writing the code and populate the data, Will this fetch Delta data for COPA datasource? for ex. if someone changes the partner function for sales order will that be captured as a delta? If not then please suggest what should be the approach for above.
    Regards,
    Girish

    Hi,
    Look up will work. it will be another way for you.
    but delta on partner function will won't work thru LO data sources.
    LO data sources are standard one and will work on their own delta function.
    You need to understand more functionally about partner function.
    this won't be the unique for one record. same function will may have multiple records.
    in such cases we don't have any option to make partner function as delta relevant field thru custom data source.
    Another work around:
    Create one custom data source on table VBPA with less fields along with partner function.
    load custom data source data to bw DSO(xyz).
    create multiprovider on custom DSO(xyz) and copa cube.
    At data source - you need to run daily full load only(info pack and dtp)
    Thanks

  • 0HR_PY_1 Datasource: Fields not filled

    Hello to everyone!
    I'm afraid that I need some help from you. We are extracting HR data via 0HR_PY_1 Datasource and I have noticed that some fields are not filled by the extraction program. I've read BW Online Help about this datasource  and I have realized that the extraction program doesn´t fill these fields.
    Do you know how can I do to make these fields to be filled?
    The fields are:
    - KOKRS_CO
    Controlling area 
    Not filled
    - CO_AREA_WP
    Controlling area of master cost center
    Not filled
    -LSTAR_CO
    Debited activity type
    Not filled
    VORNR_CO
    Debited transaction
    Not filled
    GSBER_CO
    Debited business area
    Not filled
    KOSTL_CO
    Debited cost center
    Not filled
    BUKRS_CO
    Debited company code
    Not filled
    NPLNR_CO
    Debited network
    Not filled
    AUFNR_CO
    Debited order
    Not filled
    PS_POSID_CO
    Debited WBS element
    Not filled
    Thanks in advance!
    Best regards,
    Mª Victoria Durá

    Thank you very much for your help but, I'm afraid that the procedure that you explained is for enhancing the datasource but my problem is that the fields already exist but they are not filled by extraction programmes. I don't understand why does SAP make this and I was wondering if there are any way to make these fields to be filled using the standard procedure. I mean, if these fields are not filled (and SAP tells you so in BW Online Help), why these fields are included in the extract structure and in the datasource?. I tend to think that there are any procedure to follow in order to fill these fields. What do you think about it?. In adition, the not-filled-fields are CO objects so it doesn´t appear to me to be very easy the filling of them.
    I don't know if anybody has been in the same situation or if is there any documentation about it that can help. Thank you for your help!
    Mariví

  • Multiple domains share the same datasource?

    Hi,
    I'm new to weblogic server. I understand that there's a wizard available for us to follow and have the domain created. However, does it means that only one domain can be tagged to the datasource (meaning the database used). Can I recreate or clone this admin server, yet having the same datasource in used. The reason is because I need to have another instance available for quick test and yet do not wish to spend too much time on the domain setup, etc.
    Appreciate any advise on this.
    Cheers

    You can just copy your directory and put it on another machine (for example). If you keep it on the
    same machine, you have to make sure the ports used by servers are different. Note also you
    cannot have domains of the same name on the same machine.
    When you installed the environment in production mode you have to disable hostname verification.
    As WebLogic per default uses a self-generated certificate, which has the hostname of the machine
    in it on which the domain was configured.

  • SSAS using datasource at query time

    I have a set of reports that run against the cube.  The cube has proactive caching enabled.  If the datasource within the SSAS database has invalid credentials, queries against the cube will not run.  Note: SSAS is setup to run as network
    service rather than a domain account (temporary situation).
    My understanding of MOLAP is that the data is stored within the cube an no access to the source (sql server datamart) is needed for the user to run queries against the cube.  Now this is proving to not be true... leaving me feeling like I don't know
    what is going on at all.  After 7 years using SSAS I'm finding out that the user's queries don't actually run against the cube, they require that ssas be able to connect to its source data to satisfy the query???!?!?!??!?
    I really really don't understand this, but I have tested repeatedly and when I set the datasource to invalid credentials, report queries won't run or just return 0's.  Earlier we were getting 08001 "can't connect to server" messages, now I'm
    just getting all 0's on the reports when the datasource credentials are invalid.
    Just to make this 100% clear, I am talking about the SSAS datasource within the SSAS database, not the SSRS datasource that points to the cube.  Has anyone ever seen this before?  To make it more confusing, now it seems to be working in excel regardless
    but SSRS still needs the valid datasource.
    Thanks,
    Ken

    Hi Ken,
    The data source credential of SSAS cube you specified in SSMS was use to retireve data from relational database. For example, we will need to update cube data or some MDX queries need to get data from relational DB. The Analysis Services querying architecture
    provides several components that work together to efficiently retrieve and evaluate data, please see:
    When we query a cube, the query processor breaks the query into subcube requests for the storage engine. For each subcube request, the storage engine first attempts to retrieve data from the storage engine cache. If no data is available in the cache, it
    attempts to retrieve data from an aggregation. If no aggregation is present, it must retrieve the data from the fact data from a measure group’s partition data.
    In your case, I recommend you use SQL Server Profiler to capture SSAS and Database Engine events for further investigation. If the SSRS reports need to get data from relational database, we will get detail information in this scenario.  For more information
    regardig SQL Server Profiler, please see:
    http://technet.microsoft.com/en-us/library/ms181091.aspx
    If you have any feedback on our support, please click
    here.
    Regards,
    Elvis Long
    TechNet Community Support

  • Is it possible to change the owning schema when generating datetime dimensions in the datasource?

    When creating a date time dimension using the dimension wizard and generate schema wizard, i would like to be able to change the owning schema from dbo to something else. The field is greyed out and I am unable to change it.
    Is it possible to change the owning schema when generating datetime dimensions in the datasource?

    Hi rkbjr,
    Thank you for your question.  
    I am currently looking into this issue and will give you an update as soon as possible. 
    Thank you for your understanding and support. 
    Thanks,
    Bin Long
    TechNet Subscriber Support
    If you are TechNet Subscription user and have
    any feedback on our support quality, please send your feedback here.
    Bin Long
    TechNet Community Support

  • Verifying DSN reports Datasource not found?

    Hello, thanks for looking..
    I have two datasources configured on a production web server (Win IIS, CF8), one that connects to a remote database and the other connects to a local database. A web application I was in the process of migrating to a new server used the datasource that connected to the remote database as I was taking one step at a time. Yesterday became the day to make the final jump and to migrate the database, so after comparing the two databases (SQL Delta is a beautiful tool), I proceded to swap the datasources by renaming them in the ColdFusion Administrator. I renamed the datasource to the remote database to some arbitrary backup name, then renamed the datasource to the local database to the original name of the the datasource to the remote database.
    JDBC, ODBC, whatever, I have probably performed these same steps hundreds of times without an issue. Yesterday, after renaming the datasource, verifying it, confirming it was working by browsing the application and seeing the noticable performance increase from accessing a local database, I was content to call to process complete.
    At 3am server time however, the application began throwing errors, saying the datasource could not be found. I jumped into CF Admin to check everything out and found things just as I left them, two datasources, one named as a backup and the other the original name to the local database. When I clicked verify all however, the datasource to the local database failed, saying "Datasource could not be found".
    I am curious as to what may have happened here. I don't see any cause for alarm with regards to technique as I have been using it for so long. ColdFusion server.log shows no sign of a restart. Does anyone have any clue what could have happened here? How could a datasource appear in CF Admin, yet verifying it says it doesn't exist? I resolved the problem by renaming it to something else, then renaming it back but I plan on deleting and recreating it completely this evening when traffic is lower. My only guess here is that ColdFusion updated active memory but failed when updating the registry, then something happened where things were refreshed.. But why.. and how can that be detected/avoided if true?
    I have a lot of people asking me questions now, and after spending great amount of time convincing them ColdFusion is the best answer for their needs the first time around, I am having trouble doing so this second time. So thank you in advance for any help or insight..
    Mike

    I am okay with that, and a 1 in 200ish failure rate, especially now that I have an automated script verifying datasources regularly. My clients are not however, especially without understanding the cause. Thanks for the reply though.
    Well, if that client was mine, I'd be saying something along these lines:
    "I've investigated this as much as practical, and asked around the CF community, and no-one has heard of this happening.  Given it's an isolated occurrence, I'm prepared to put this down to a temporary aberration, which is now resolved.  "Case closed", although obviously we'll "watch this space" too.  All computer software is very complex, and there will always be glitches.  I can investigate this to conclusion if you like, but it's difficult to troubleshoot something after havnig fixed it, especially when it's not replicable.  I am satisfied this is not something to be concerned about, however if you really want me to get to the bottom of it, it's likely to take a week (billable, as it's not a good use of my time), and at the end of it, you're not going to be any better off.  If it's such a concern - and fair enough - that there's a single point of failure here, then let's start talking about some fail-over protection.  That way if any other aberrations present themselves, it will not be an issue".
    However, in the mean time, have a look at your JRun and CF logs at round about the time the problem happened, and see if there's anything in there to do with JDBC and the like.
    Adam

  • Invalid extract structure template && of DataSource &&

    Hi,
    While creating a z datasource i have received the above error. Fom the previous threads here (and from notes 202162 & 335342) I can understand that when there are Currency or Quantity fields included in the datasource, their key fields (CUKY for currency and UNIT for Quantity) shoul also be included in the same view and structure.
    In this case I am taking data in the view from table EKBE. I need to have MENGE & DMBTR. For this I am also taking fields LSMEH (Datatype - Unit) and WAERS (Datatype - CUKY) respectively. Yet the error is coming.
    If i try to join EKPO table in the view then this error does not come as I give EKPO-MEINS as ref table and ref field for MENGE. This thing works even if datatype of EKPO-MEINS is also UNIT Datatype. BUt I can not use join on EKPO table as while extracting the data in RSA3 it is giving me data from both tables (which is not the requirement)
    Plz help.

    Hi,
    Try giving EKPO-MEINS as ref table and ref field for MENGE. without joining EKPO.
    Hope it helps
    Regards,
    Rathy

  • How to insert a null value to combobox which is bind with datasource

    Hi,
    i am working on c# winforms application. I have 2 comboboxes one is CustomerCbBox and another is OrderCbBox. CustomerCbBox is bind with data source.
    on CustomerCbBox SelectedIndexChanged event data is populated in OrderCbBox from data source. 
    i want to add a null or empty field in CustomerCbBox. all I want is that when the user don't Select Customer Name and OrderNo null or empty data enter in database. but when i click Insert Button it generate error "Object reference is not "object
    reference not set to an instance of an object".
    here is my code.
      private void MainForm_Load(object sender, EventArgs e)
    production pd = new production();
    CustomerCbBox.DataSource = pd.Customer();
    CustomerCbBox.DisplayMember = "Cust";
    CustomerCbBox.ValueMember = "CustId";
     private void CustomerCbBox_SelectedIndexChanged(object sender, EventArgs e)
                      try
                          string PartyName = CustomerCbBox.SelectedIndex.ToString();
                           PONum1.DataSource = pd.GetPOnumActive(PartyName1);
                              PONum1.DisplayMember = "PONum";
                              PONum1.ValueMember = "PONum";
                      catch (Exception ex)
                          MessageBox.Show(ex.Message);
         Please help me to solve this problem thanks.      

    Hello,
    We could check whether the user has selected any items before adding that to database.
    if (CustomerCbBox.SelectedIndex > -1)
    // do something
    }else{//add dbnull to database}
    And for adding null to databse, we could pass DBNull.Value instead.
    Regards,
    Carl
    We are trying to better understand customer views on social support experience, so your participation in this interview project would be greatly appreciated if you have time. Thanks for helping make community forums a great place.
    Click
    HERE to participate the survey.

  • Error while changing the physical IP address of SOA related datasource

    Hello Everyone!!!
    I am facing a problem while changing the ip address of my SOA related datasource in weblogic 10.3,Below is a brief description about my prb:
    I have created a domain in weblogic 10.3 and have configured a datasource to it and its working fine initially.
    But i had to change the ip address of my datasource(i.e Edit the existing the setting of my domain) as my database was migrated to another m/c.The changes that i did was,I changed the url in the connectionpool tab of each of the datasource's listed in admin console(Please let me know if i have left out any other important place in which i had to make this change).
    After making this change my ManagedServer starts but my Enterprise Application named soa-infra doesnt start and when i checked the log to find out the error,this is what i came across:
    B2B-50025: Repository error: Error -: B2B-51572: Error reading input stream.: java.io.IOException: Stream closed
         at oracle.tip.b2b.init.Repository.initExchange(Repository.java:658)
         at oracle.tip.b2b.init.Repository.initialize(Repository.java:764)
         at oracle.tip.b2b.init.B2BServer.readRepository(B2BServer.java:509)
         at oracle.tip.b2b.init.B2BServer.initialize(B2BServer.java:212)
         at oracle.tip.b2b.engine.ThreadWorkExecutor.startB2B(ThreadWorkExecutor.java:345)
         at oracle.tip.b2b.engine.ThreadWorkExecutor.run(ThreadWorkExecutor.java:203)
         at oracle.integration.platform.blocks.executor.WorkManagerExecutor$1.run(WorkManagerExecutor.java:105)
         at weblogic.work.j2ee.J2EEWorkManager$WorkWithListener.run(J2EEWorkManager.java:183)
         at weblogic.work.DaemonWorkThread.run(DaemonWorkThread.java:30)
    Caused by: Error -: B2B-51572: Error reading input stream.: java.io.IOException: Stream closed
    Any suggestion would be of great help.
    Thks in advance!!!

    Hey thks for the reply,
    We were able to solve the problem by just restarting our database,
    i got some reference from the below link(pls refer to this topic:Changing the IP Address of a Metadata Repository Installation):
    [http://download.oracle.com/docs/cd/E12839_01/core.1111/e10105/host.htm#CHDHAEFE]
    but m still confused do i have to restart my database server each time i change the DB server that m pointing to in my weblogic console and if yes can you pls help me understand what change happens when the database restarts that makes this work.
    Another issue that m facing now is that after solving the above prb(by restarting the DB) m not able to deploy process's on to my server.
    The Exception that appears in the log file is as follows:
    [2011-02-08T13:03:33.220+05:30] [soa_server1] [ERROR] [] [oracle.soa.bpel.engine.compiler] [tid: [ACTIVE].ExecuteThread: '17' for queue: 'weblogic.kernel.Default (self-tuning)'] [userId: weblogic] [ecid: 0000Is4C5SLBLA1LzuP5iW1DKEXf000d33,0] [APP: soa-infra] [dcid: d61ccdb1526887f2:678e40b9:12e040e4cba:-7ffd-000000000000290f] <BPEL1Processor::generateProcessClass> MerchantSales_BPELProcess.bpel(line 228): invalid query[[
    query "/EmailPayload/ns3:FromAccountName" is invalid, because java.lang.IllegalArgumentException: null schema location for ns=http://xmlns.oracle.com/ias/pcbpel/NotificationService [Cause=null schema location for ns=http://xmlns.oracle.com/ias/pcbpel/NotificationService]
    Check the XML schema and make sure your query string is valid
    [2011-02-08T13:03:33.448+05:30] [soa_server1] [ERROR] [] [oracle.soa.bpel.engine.deployment] [tid: [ACTIVE].ExecuteThread: '17' for queue: 'weblogic.kernel.Default (self-tuning)'] [userId: weblogic] [ecid: 0000Is4C5SLBLA1LzuP5iW1DKEXf000d33,0] [APP: soa-infra] [dcid: d61ccdb1526887f2:678e40b9:12e040e4cba:-7ffd-000000000000290f] <CubeProcessLoader::create> [[
    java.lang.Exception: BPEL compilation failed
         at com.collaxa.cube.lang.compiler.BPEL1Processor.generateProcessClass(BPEL1Processor.java:529)
         at com.collaxa.cube.lang.compiler.BPEL1Processor.loadProcess(BPEL1Processor.java:495)
         at com.collaxa.cube.lang.compiler.BPEL1Processor.loadProcess(BPEL1Processor.java:438)
         at com.collaxa.cube.lang.compiler.BPEL1Processor.createCubeProcess(BPEL1Processor.java:396)
         at com.collaxa.cube.engine.deployment.CubeProcessFactory.create(CubeProcessFactory.java:61)
         at com.collaxa.cube.engine.deployment.CubeProcessLoader.create(CubeProcessLoader.java:112)
         at com.collaxa.cube.engine.deployment.DeploymentManager.deployProcess(DeploymentManager.java:238)
         at com.collaxa.cube.engine.deployment.DeploymentManager.deployComponent(DeploymentManager.java:182)
         at com.collaxa.cube.ejb.impl.CubeServerManagerBean._deployOrLoadComponent(CubeServerManagerBean.java:726)
         at com.collaxa.cube.ejb.impl.CubeServerManagerBean.deployComponent(CubeServerManagerBean.java:114)
         at com.collaxa.cube.ejb.impl.bpel.BPELServerManagerBean.deployComponent(BPELServerManagerBean.java:85)
         at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
         at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
         at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
         at java.lang.reflect.Method.invoke(Method.java:597)
         at com.bea.core.repackaged.springframework.aop.support.AopUtils.invokeJoinpointUsingReflection(AopUtils.java:310)
         at com.bea.core.repackaged.springframework.aop.framework.ReflectiveMethodInvocation.invokeJoinpoint(ReflectiveMethodInvocation.java:182)
         at com.bea.core.repackaged.springframework.aop.framework.ReflectiveMethodInvocation.proceed(ReflectiveMethodInvocation.java:149)

  • Error while invoking bean "finder": Instance not found in datasource

    Hi All
    I am facing Error while trying to execute a BPEL processing using Notification service for sending out mail .
    Following is the Error Stack :
    <ERROR> <default.collaxa.cube> <BaseCubeSessionBean::logError> Error while invoking bean "finder": Instance not found in datasource.
    The process domain was unable to fetch the instance with key "677c0310e56e2b26:26d607:10dca65e80a:-7ffc" from the datasource.
    Please check that the instance key "677c0310e56e2b26:26d607:10dca65e80a:-7ffc" refers to a valid instance that has been started and not removed from the process domain.
    ORABPEL-02152
    Instance not found in datasource.
    The process domain was unable to fetch the instance with key "677c0310e56e2b26:26d607:10dca65e80a:-7ffc" from the datasource.
    Please check that the instance key "677c0310e56e2b26:26d607:10dca65e80a:-7ffc" refers to a valid instance that has been started and not removed from the process domain.
         at com.collaxa.cube.engine.data.AdminFinder.__lookupInstanceMetaData(AdminFinder.java:314)
         at com.collaxa.cube.engine.data.AdminFinder.__lookupInstanceMetaData(AdminFinder.java:262)
         at com.collaxa.cube.engine.data.AdminFinder.lookupInstanceByConversationId(AdminFinder.java:235)
         at [b]com.collaxa.cube.ejb.impl.FinderBean.lookupInstanceByConversationId(FinderBean.java:177)
         at IFinderBean_StatelessSessionBeanWrapper42.lookupInstanceByConversationId(IFinderBean_StatelessSessionBeanWrapper42.java:669)
         at com.oracle.bpel.client.Locator.lookupInstance(Locator.java:420)
         at displayInstance.jspService(_displayInstance.java:111)
         at com.orionserver.http.OrionHttpJspPage.service(OrionHttpJspPage.java:56)
         at oracle.jsp.runtimev2.JspPageTable.service(JspPageTable.java:350)
         at oracle.jsp.runtimev2.JspServlet.internalService(JspServlet.java:509)
         at oracle.jsp.runtimev2.JspServlet.service(JspServlet.java:413)
         at javax.servlet.http.HttpServlet.service(HttpServlet.java:853)
         at com.evermind.server.http.ServletRequestDispatcher.invoke(ServletRequestDispatcher.java:824)
         at com.evermind.server.http.ServletRequestDispatcher.forwardInternal(ServletRequestDispatcher.java:330)
         at com.evermind.server.http.ServletRequestDispatcher.forward(ServletRequestDispatcher.java:222)
         at com.collaxa.cube.fe.DomainFilter.doFilter(DomainFilter.java:152)
         at com.evermind.server.http.ServletRequestDispatcher.invoke(ServletRequestDispatcher.java:663)
         at com.evermind.server.http.ServletRequestDispatcher.forwardInternal(ServletRequestDispatcher.java:330)
         at com.evermind.server.http.HttpRequestHandler.processRequest(HttpRequestHandler.java:830)
         at com.evermind.server.http.HttpRequestHandler.run(HttpRequestHandler.java:285)
         at com.evermind.server.http.HttpRequestHandler.run(HttpRequestHandler.java:126)
         at com.evermind.util.ReleasableResourcePooledExecutor$MyWorker.run(ReleasableResourcePooledExecutor.java:192)
         at java.lang.Thread.run(Thread.java:534)
    <2006-09-20 14:55:44,109> <ERROR> <default.collaxa.cube.services> <PCException::<init>> Error in sending e-mail message.
    <2006-09-20 14:55:44,119> <ERROR> <default.collaxa.cube.services> <PCException::<init>> Error in sending e-mail message.
    The understanding formed till now is that the service is trigged using a stateless session bean wrapper and the lookup for the same is failing.
    One option will be to write the java code for performing the lookup using method
    lookupInstanceByConversationId(String convId, DomainAuth auth)
    but certain information e.g DomainAuth is not available.
    I have tried locating information about the dynamic wrapper which is being created.
    Any inputs/comments/direction for the above problem are welcomed
    Cheers

    Thanks Marc.
    I checked the domain log (ORA_BPEL_HOME\bpel\domains\default\logs\domain.log) for information. It repeats the same error displayed in the bpel server window.
    I have now put some SOPs in my Java embedding activity to see if the process gets instantiated or not. On the BPEL server window, I am able to see these SOPs upto the point where I invoke a method invocation on my remote Java class. Which clearly means the BPEL process is instantiated, contrary to the information in the error (that the BPEL instance cannot be found) I am getting.

Maybe you are looking for

  • Counter in file name in receiver file adapter

    Hi, I am using FTP, my requirement is to add counter in the file name i.e XYZ001. And added to this i can increment the counter till XYZ020, after this i am supposed to start again from XYZ001.It will be some how handled that when i place XYZ001 for

  • Need help in connecting Dell Monitor to Dell Inspiron

    I recently purchased an Inspiron 17R laptop to reduce the high usage of my Precision M6600 laptop. A Dell U2312HM monitor was used with the Precision as a second monitor and I want to use it with the newly acquired Inspiron system as a second monitor

  • SVG files?

    What are SVG files used for? Are they still used? Thanks.

  • How to update resource content (XML Versioning)

    Hi, How to update resource content in the resource_view? I am trying to use XML Versioning. And not able to update the resource content using PL/SQL. Upon executing update, sql prompt says 1 row updated, but when I extract the same resource, it retur

  • Pse elements 7 missing standard menu bar

    I am using pse elements 7. I have reloaded the program onto a new harddrive and my old one crashed. The pse 7 loads onto the new hardrive but the standard menu bar ( the tool bar at the top of the screen) does show up. The one with file, edit, find,