Data source migration in parallel to CRM server upgradation

Hi Experts,
We are working on BI 7.0 system which pulls data from CRM 2004. Recently we upgraded BI environment from BW3.5 (to BI.70 SP -16).
Now we are re-modeling the existing BI design due to performance issues & to use BI functionalities. We would like to migrate the data sources in BI also to BI version so that we can get advantage of using DTP & transformations.
But the issue here is now we started to up grade our source system also from CRM 2004 --> CRM 2007. I beleive that the data source version will be maintained in source system (CRM) tables also.
So if we migrate the data source now (before CRM upgrade), will it cause any issues in BI once CRM server upgrade done?
Note: We can implement all of our new designs with out data source migration.
Please advice us to proceed. Let me know if you need more details.
Thanks in advance.
Regards,
Arun Thangaraj.

Hi reddy,
1). Why Migration Process?
Migration process is to convert BW 3.x objects ( such as Data source, Trasfer rules and update rules etc) into BI 7.0 objests.
2). Why we are Migrating BW 3.X objects into 7.0 objects?
We are Migrating 3.x data source into 7.0 inorder to avail three functions, Those are
a). Direct data acquisition
b). Realtime Data Acquisition (RDA)
c). BI Accelarator.
3) .Regarding With export and Withour export options ?
These two options come in the process of Migration step. Regarding these options clearly described in the fourth point as bellow.
4). What will happen in this Migration process?
( Remember at the time Migration Transfer structure and Transfer rlues deleted, keep in mind go throgh bellow steps).
What ever may be the option selected (i.e, With Export or Without export ) data source will migrate to 7.0.
But if you choose With export option deleted Transfer Structure and Trasfer Rules will store in the table " RSDSEXPORT". If you select other option (i.e, Without export ) the deleted Trasfer structure and trasfer rules perminently not avaliable.
Thats why always go with With Export option only. This will help to revert back 7.0 data source into 3.5 data source when ever is there any requirement in future.
Please go look into bellow link.
Re: Migration of datasource from 3.5 to 7.0
if u r satisfed about my answer please assign points.
Thanks & Regards,
Kotesh.

Similar Messages

  • BOE XI v3.1 Data Source Migration Wizard

    The BOE XI v3.1 Data Source Migration Wizard is crashing after trying to convert 500 .rpt files.
    The Error message is below. We have tried this a few times now without success.
    "The application, C:\Program Files\Business Objects\BusinessObjects Enterprise 12.0\win32_x86\BVMigrationWiz.exe, generated an application error The error occurred on 02/24/2009 @ 18:04:33.467 The exception generated was c0000005 at address 3784622E (CRPE32!Ordinal995)"
    We are running Windows 2003 sp2

    I'm getting in eventviewer:
    Faulting application BVMigrationWiz.exe, version 12.1.0.882, faulting module crpe32.dll, version 12.3.0.601, fault address 0x0001625e.

  • How to set up data source for JDeveloper's embedded OC4J Server?

    Help needed. I Created a BMP EJB and tried to test
    it with JDeveloper's( version 9.0.3) embedded
    OC4J server. For JDBC, I created a data source
    in data-source.xml lacated in JdeveloperHome\j2ee\home\config. However,
    when I run the test client. I got the error
    from JNDI lookup:
    java.rmi.RemoteException: jdbc/entdatasource not found
    Could anybody with similar experience help?
    Thank you in advance.
    -John Wang

    Try making your changes in the data-source.xml located @ JDev_root\jdev\system9.0.3.1035\oc4j-config.
    You can tell that this is the one to be modified by watching the log when you start the OC4J server. Look for the line that says C:\JDEV\JDK\bin\javaw.exe -ojvm -classpath . . .
    the -config option will point to where the server.xml is used from, the same dir should be where you get your data-source.xml

  • ODBC data source migrations

    How do you move ODBC system DSN from one server to another?  There are about 35 ODBC connections on the server that I need on the new server since there are reports that use them.  I do not think the Import Wizard will do this... thanks!

    This is a free tool that should allow you to export your DSN's to a file and import to a different machine.  As this isn't an SAP written tool, use at your own risk.  The company reports that it supports all the major Windows OS versions.
    http://www.sqledit.com/odbc/

  • How to make CRM data sources for delta capability?

    Hi All,
    I know that the CRM data sources 0CRM_SALES_ACT_1, 0CRM_QUOTATION_I, 0CRM_SALES_ORDER_I have delta capability.
    At present we are doing full loads daily but it has got some performance issues.So we are thinking to make these as delta loads.
    Already some one has made a trail to make delta loads but not successful for some unknown reasons.
    Could you please suggest me the steps to make these data sources delta capable both in CRM and BW systems?
    As this is urgent...Please suggest ASAP.

    Hi
    Go thorugh Below the note: 692195
    Summary
    Symptom
    There may be problems or issues related to data tranfer from CRM to BW.
    Other terms
    CRM-BW extraction,upload,initial,delta,full upload ,Sales Analytics,
    Reason and Prerequisites
    There could be errors in customization or program errors due to which
    data may not be transferred or incorrectly transferred to BW.
    Solutions
    Question 1 : The Extraction from CRM to BW takes a very long time. What can be done? (Performance Issues)
    Suggestion 1: Please implement notes  653645 (Collective note) and
    639072(Parallel  Processing).
    The performance could be slow because of the wrong control parameters
    used for packaging.
    You can change the package size for the data extraction.
    Also note that changing the package size in the transaction SBIW
    would imply a change for all the extractors. Instead, you could
    follow the path in the bw system.
    Infopackage (scheduler)    > Menu 'Scheduler'   > 'DataS. default data
    transfer'   > maintain the value as 1500 or 1000(This value is variable)
    The package size depends on the Resources available at the customer side
    (The no of parallel processes that could be assigned =
    1.5 times the no of CPU's available approx.)
    Question 2 : On executing transaction RSA3 I get records but I find 0
    records when I load data from BW request.(No Data Available)
    Suggestion 2: First check if there are any entries in the table
    CRMD_ORDER_INDEX.Only if there are entries in this table you can
    extract records.
    If this is not the case then,
    It is possible that the user does not have sufficient
    authorities for extraction of  the relevant objects.
    Additionally, please review and implement the following notes
    615670
    161570
    150315
    618953
    If you are in the release 4.0 then
    To do BW extraction with the user please see that the following
    authorization object exists(display mode is enough):
    CRM_ACT,CRM_OPP, CRM_LEAD, CRM_SAO, CRM_SEO, CRM_CO_SE,CRM_CO_SC
    CRM_CO_SA, CRM_CON_SE, CRM_CMP, CRM_ORD_OP,CRM_ORD_LP,CRM_ORD_PR
    CRM_ORD_OE, CRM_CO_PU, CRM_CO_PD, CRM_ORD_PO
    (all these objects are linked to transaction crmd_order).
    Question 3 : The Deltas for my data source are not extracted . What can I do?
    Suggestion 3: Please check the following.
    Please Check if the services have been generated in transaction GNRWB.
    If they are not active(not marked 'X' before their names) then activate
    the services following the steps here.
    Go to transaction GNRWB
    Select BUS_TRANS_MSG
    Select (on the right, the services) : BWA_DELTA3, BWA_FILL, BWA_queue
    Press Generate.
    Also check  for the following:
    1. The delta should have been initialized successfully.
    2. Confirm that all Bdocs of type BUS_TRANS_MSG
       are processed with success in SMW01.
    3. If there are queues in SMQ1 with erroneous status then activate
       these queues.
       In Transaction SMQ1 if there are Queues existing with
       names beginning with CRM_BWAn (n is number) then
       activate these queues in the same transaction.
    4.a)If required activate the datasource
        Go to transaction BWA5   > select the required datasource and
        activate.
    4 b) The Delta may not be active ,activate the delta in BWA7 by
      selecting the name of the datsource and pressing the candle icon for
      'activate delta'.
    5. In BW system
       Go to transaction RSA1   > modeling   > infosources   > select the
       infosource   > right mouse click on the selected
       infosource   > choose option replicate datasource
        Activate the infosource.
    6. Go to the scheduler for the infosource   > select delta  in the
        update  >choose the option PSA only (in the Processing tab)
        > start immediately
    Check the entry in the RSA7 in the OLTP(CRM system)
    Question 4: How can I extract the fields, which are not provided in the standard  data source extraction .
    Suggestion 4: Follow the steps mentioned below.
    1. Enhance Extract Structure with the required fields. (Create & include
       an append structure to the extract structure via transaction RSA6).
    2.a) Release the fields of the append  for usage. (  To do this, double
       click on the Datasource and remove the flags in the column 'Hide
       Field' for all fields of Append. )
    2.b)If the new fields cannot be seen in the extract structure of
        the transaction BWA1 then change and save the datasource, and then
        activate it in RSA6.
    3. Define your mappings in BADI (CRM_BWA_MFLOW) to fill these fields.
    Goto SPRO .
    Follow the path ->
    SAP Implementation guide ->Implementation with other mySAP components ->
    Data transfer to the Business Information Warehouse->
    Settings for the application specific datasources (CRM)->
    Settings for BW adapter->
    Badi :BW adapter :Enhancement of datasources in messaging flow.
    4. Replicate the new Datasource to BW.
    5. Expand the Communication Structure in BW.
    6. Maintain transfer Rules for the new Datasource.
    7. Activate the trasfer rules and perform the upload.
    Question 5:  I am unable to extract  user status correctly.What should I
    do?
    Suggestion 5 :Check the following notes
    531875
    616062
    713458
    700714
    765281
    Question 6 : What can I  do when the activity/Opportunity/Complaint
    reasons(Code,CodeGruppe,Katalogart)  are not extracted.
    Suggestion 6 :Check the following notes
    481686
    516820
    603609
    617411
    711146
    Question 7:Deleted opportunities are not reflected in BW.
    Suggestion 7: Check the note 706327.
    Question 8: How do you  activate the metadata?
    Suggestion 8:CRM BW adapter meta data has to be activated first before
    it is available in the system. You can use Transaction BWA5 to copy the
    meta data for selected DataSources. You can reach the transaction via
    the IMG maintenance 'SAP Reference IMG -> Settings for SAP Business
    Information Warehouse -> Activate BW Adapter Meta Data'. For more
    information, see the documentation on the IMG activity 'Activate BW
    Adapter Meta Data'.  (Note 432485)
    Question 9: I donot get any records for the delta upload of my attribute
    datasource(s).What is to be done?
    Suggestion 9: In case of attribute datasources, it is possible that the
    entry for the GUID is missing in the table SMOXAFLD.
    If , for example the datasource 0CRM_OPPT_ATTR is not giving deltas
    then you can follow the steps:
    1)If Delta process is active for the attribute datasource e.g.
         0CRM_OPPT_ATTR,  then stop the delta process in the BW
         system 
    2) In the CRM system, Make the entry in the table SMOXAFLD
       for the datasource with the Key
       as  0CRM_OPPT_ATTR     GUID 3) Save the entry.Activate the datasource                           4) Check that the above entry is replicated in smoxafld_s also    after this.               5) Create a transport request manually for the following    object             R3TR   SMO4   0CRM_OPPT_ATTR    
    Question 10: What do the status BWSTONESYS0 , BWSTONEUSS0,BWSTTECSYS0 and various other BW status mean ?
    Suggestion 10: The BW status are used to extract system and user defined
    status.
    The BW status are defined in the customization settings in SPRO.
    Check for-> Status Concept for BP/Product/CRM objects
    Here goto-> Process user status You will find the documentation attached here for the user status. Going inside the transaction you will find the status groups USS0, ZIOP,ZMOP etc. The names of the various BW status are derived from this For ex. BW + ST+ One + USS0 gives the name of the field BWSTONEUSS0 or BWST + ONE + ZMOP = BWSTONEZMOP (Master opportunity values) (which means BWST + (status group name) + status object group name) Double clicking on any of the object groups will take you to the values that these status can have . For ex. BWSTONEUSS0 in your system can have values E001 ,E002,E003,E007 which will be shown in RSA3 as BW status values 1,2,3,7, respectively.
    Similarly we have Goto -> Process system status (in SPRO). Here you can get the values for the system status in exactly the same way as BWSTONESYS0(Lifecycle status) , BWSTTECSYS2 (Error) etc. In RSA3 you get the names as BW status, To know which corresponds to which status here,Gotothe record list in RSA3 . Here goto Settings - Layout -Current . Right click -> Press Show technical field names . You will be able to see the BW status names and will be able to adjust the layout accordingly.
    Thank you,
    DST

  • 10.1.3 Migration Issue with data-sources.xml

    Hi:
    Running into lots of migration issues for an application built in 10.1.3 EA, migrating to 10.1.3. production, mostly in the BC project. Here's one: the database connections built in EA were not carried over to prod, so they needed to get re-built. No biggie. But now the data-sources.xml file in the model project has duplicate entries for my data sources. I have tried directly editting that file and removing the duplicates, but when I go to run my application (ADF Faces & ADF BC) JDeveloper hangs. When I abort JDev and re-start it, the duplicate entries are back in data-sources.xml (and yes, I did save the file after making the changes). It seems that the xml file is built dynamically from definitions stored elsewhere. Where is the source? What will I need to modify to get the changes to take?
    Thanks.
    Johnny Lee

    There is new control over when/whether JDeveloper add/updates your data-sources.xml file for the embedded server. The settings are located here:
    Tools | Embedded OC4J Server Preferences | Current Workspace | Data Sources
    There are some settings there that control whether/when we update the data-sources.xml

  • While creating the Data source for Sql server am getting this error?

    Hi i new to Power BI sites and office 365,
    For my first step 
    1. i have successful creaeted Gateway in Office 365
    2.Now i have creating  the Data source for that gateway using Sql server R2
    3.While Creating the Data source using the Corresponding gateway i have given and next phase set Credentials  phase i have noticed one pop up window and it will shows me like
    Data source Settings and  below in that window it will show's u the things like 
    Datasource Name.
    Loading......
    Credentials type:
    Privacy Level:
    Failed to verify parameter
    I will get this results finally  ,so hw should i achive this probelm ,Please let me know if any one knows
    So kindly give me the proper answer for this.
    Regards

    Any suggestions for Chinnarianu?
    Thanks!
    Ed Price, Azure & Power BI Customer Program Manager (Blog,
    Small Basic,
    Wiki Ninjas,
    Wiki)
    Answer an interesting question?
    Create a wiki article about it!

  • Connection/data source caching in JSC App Server?

    Tried to change the data source for one of my apps (changed all data sources on all row sets). I got a connection error despite the fact that the IDE could see the data source.
    I restarted the app server to no avail.
    I finally stopped the app server AND restarted JSC and eveything started working again (except for the Conversion error I'm tracking down :-) ...

    Did the connection error occur at runtime ?
    If so, then it's not recommended to change the datasource
    especially if you change databases. It's not impossible to make
    the changes - you need to edit the Page bean constructor and the
    constructor in the SessionBean and anywhere else you
    initialize the rowset with the URL.
    If you changed the database, then because vendors
    map JDBC types differently, you may need to make
    changes . Typically you'll see conversion errors when changing
    the database.
    JDBC Type (which we call SQL Type) --> Java Type:
    =========================================
    CHAR --> java.lang.String
    VARCHAR --> java.lang.String
    LONGVARCHAR --> java.lang.String
    NUMERIC --> java.math.BigDecimal
    DECIMAL --> java.math.BigDecimal
    BIT --> boolean
    BOOLEAN --> boolean
    TINYINT --> byte
    SMALLINT --> short
    INTEGER --> int
    BIGINT --> long
    REAL --> float
    FLOAT --> double
    DOUBLE --> double
    BINARY --> byte[]
    VARBINARY --> byte[]
    LONGVARBINARY --> byte[]
    DATE --> java.sql.Date
    TIME --> java.sql.Time
    TIMESTAMP --> java.sql.Timestamp
    CLOB --> java.sql.Clob
    BLOB --> java.sql.Blob
    ARRAY --> java.sql.Array
    DISTINCT --> (mapping of underlying type)
    STRUCT --> java.sql.Struct
    REF --> java.sql.Ref
    DATALINK --> java.net.URL
    JAVA_OBJECT --> (underlying Java class)
    Source: JDBC 3.0 Specification, Appendix B
    Also, see this post for more info :
    http://swforum.sun.com/jive/thread.jspa?forumID=123&threadID=47015
    John
    JSC QA

  • OPA Data Source Connector failed to run on Linux server

    When I run data source connector for java on linux server using:
    java -Xms512m -Xmx512m -classpath "/usrfile/OPA/DataSourceConnector/lib/determinations-engine.jar:/usrfile/OPA/DataSourceConnector/lib/log4j-1.2.15.jar:/usrfile/OPA/DataSourceConnector/lib/sqlitejdbc-v044-native.jar:/usrfile/OPA/DataSourceConnector/lib/determinations-data-source-connector.jar:/usrfile/OPA/DataSourceConnector/lib/jsr173_api.jar:/usrfile/OPA/DataSourceConnectorlib/sjsxp.jar" com.oracle.determinations.connector.DataSourceConnector /usrfile/OPA/DataSourceConnector/data/DEMO3/run/DEMO3.xml -debug
    I got the following error message.So I think the path “..\conf\configuration.xsd” is hard coded and it raised this error. The detail error log is:
    27-12-2010, 11:35:20, INFO , Processing file /usrfile/OPA/DataSourceConnector/data/DEMO3/run/DEMO3.xml.
    27-12-2010, 11:35:20, ERROR, /usrfile/OPA/DataSourceConnector/bin/..\conf\configuration.xsd (No such file or directory)
    org.xml.sax.SAXException: /usrfile/OPA/DataSourceConnector/bin/..\conf\configuration.xsd (No such file or directory)
    at gnu.xml.validation.xmlschema.XMLSchemaSchemaFactory.newSchema(libgcj.so.7rh)
    at javax.xml.validation.SchemaFactory.newSchema(libgcj.so.7rh)
    at javax.xml.validation.SchemaFactory.newSchema(libgcj.so.7rh)
    at com.oracle.determinations.connector.schema.validation.Validate.compileSchema(Unknown Source)
    at com.oracle.determinations.connector.schema.validation.ValidateSAXStream.<init>(Unknown Source)
    at com.oracle.determinations.connector.XMLTools.validateXML(Unknown Source)
    at com.oracle.determinations.connector.Configuration.<init>(Unknown Source)
    at com.oracle.determinations.connector.DataSourceConnector.kickOff(Unknown Source)
    at com.oracle.determinations.connector.DataSourceConnector.main(Unknown Source)
    Caused by: java.io.FileNotFoundException: /usrfile/OPA/DataSourceConnector/bin/..\conf\configuration.xsd (No such file or directory)
    at gnu.java.nio.channels.FileChannelImpl.open(libgcj.so.7rh)
    at gnu.java.nio.channels.FileChannelImpl.<init>(libgcj.so.7rh)
    at gnu.java.nio.channels.FileChannelImpl.create(libgcj.so.7rh)
    at java.io.FileInputStream.<init>(libgcj.so.7rh)
    at gnu.java.net.protocol.file.Connection.connect(libgcj.so.7rh)
    at gnu.java.net.protocol.file.Connection.getInputStream(libgcj.so.7rh)
    at java.net.URL.openStream(libgcj.so.7rh)
    at gnu.xml.validation.xmlschema.XMLSchemaSchemaFactory.getDocument(libgcj.so.7rh)
    at gnu.xml.validation.xmlschema.XMLSchemaSchemaFactory.newSchema(libgcj.so.7rh)
    ...8 more
    I have checked the file path using following command:
    [root@bjx4 ~]# more /usrfile/OPA/DataSourceConnector/bin/..\conf\configuration.xsd
    /usrfile/OPA/DataSourceConnector/bin/..confconfiguration.xsd: No such file or directory
    [root@bjx4 ~]# more /usrfile/OPA/DataSourceConnector/bin/../conf/configuration.xsd
    <?xml version="1.0" encoding="UTF-8"?>
    <xsd:schema xmlns:xsd="http://www.w3.org/2001/XMLSchema" xmlns="http://oracle.co
    m/determinations/connector/data-source" targetNamespace="http://oracle.com/deter
    minations/connector/data-source" elementFormDefault="qualified" version="10.0.0:
    20091013" xml:lang="en">
    <xsd:element name="configuration" type="configuration-type"/>
    <xsd:complexType name="configuration-type">
    <xsd:sequence>
    <xsd:element name="threads" type="threads-type" minOccurs="0" maxOccurs
    ="1"/>
    <xsd:element name="run-limit" type="run-limit-type" minOccurs="0" maxOc
    curs="1"/>
    <xsd:element name="time-out" type="time-out-type" minOccurs="0" maxOccu
    rs="1"/>
    <xsd:element name="unknown-value" type="unknown-type" minOccurs="0" max
    Occurs="1"/>
    <xsd:element name="uncertain-value" type="uncertain-type" minOccurs="0"
    maxOccurs="1"/>
    <!-- required elements -->
    <xsd:element name="data-sources" type="data-sources-type" minOccurs="1"
    maxOccurs="1"/>
    <xsd:element name="data-mappings" type="data-mappings-type" minOccurs="
    1" maxOccurs="1"/>
    [root@bjx4 ~]# more /usrfile/OPA/DataSourceConnector/bin/..\conf\configuration.xsd
    /usrfile/OPA/DataSourceConnector/bin/..confconfiguration.xsd: No such file or directory
    [root@bjx4 ~]# cls
    -bash: cls: command not found
    [root@bjx4 ~]# clear
    [root@bjx4 ~]# more /usrfile/OPA/DataSourceConnector/bin/..\conf\configuration.xsd
    /usrfile/OPA/DataSourceConnector/bin/..confconfiguration.xsd: No such file or directory
    [root@bjx4 ~]# more /usrfile/OPA/DataSourceConnector/bin/../conf/configuration.xsd
    <?xml version="1.0" encoding="UTF-8"?>
    <xsd:schema xmlns:xsd="http://www.w3.org/2001/XMLSchema" xmlns="http://oracle.com/determinations/connector/data-source" targetNamespace="http://oracle.com/determinations/connector/data-source" elementF
    ormDefault="qualified" version="10.0.0:20091013" xml:lang="en">
    <xsd:element name="configuration" type="configuration-type"/>
    <xsd:complexType name="configuration-type">
    <xsd:sequence>
    <xsd:element name="threads" type="threads-type" minOccurs="0" maxOccurs="1"/>
    <xsd:element name="run-limit" type="run-limit-type" minOccurs="0" maxOccurs="1"/>
    <xsd:element name="time-out" type="time-out-type" minOccurs="0" maxOccurs="1"/>
    <xsd:element name="unknown-value" type="unknown-type" minOccurs="0" maxOccurs="1"/>
    <xsd:element name="uncertain-value" type="uncertain-type" minOccurs="0" maxOccurs="1"/>
    <!-- required elements -->
    <xsd:element name="data-sources" type="data-sources-type" minOccurs="1" maxOccurs="1"/>
    <xsd:element name="data-mappings" type="data-mappings-type" minOccurs="1" maxOccurs="1"/>
    <xsd:element name="output" type="output-type" minOccurs="1" maxOccurs="1"/>
    </xsd:sequence>
    </xsd:complexType>
    <!-- threads -->
    <xsd:complexType name="threads-type">
    <xsd:attribute name="value" type="xsd:integer" use="required">
    </xsd:attribute>
    </xsd:complexType>
    <!-- run-limit -->
    <xsd:complexType name="run-limit-type">
    <xsd:attribute name="value" type="xsd:integer" use="required">
    </xsd:attribute>
    </xsd:complexType>
    <!-- time-out -->
    <xsd:complexType name="time-out-type">
    <xsd:attribute name="value" type="xsd:integer" use="required">
    </xsd:attribute>
    </xsd:complexType>
    <!-- unknown value -->
    <xsd:complexType name="unknown-type">
    <xsd:attribute name="value" type="xsd:string" use="required">
    </xsd:attribute>
    </xsd:complexType>
    <!-- uncertain value -->
    <xsd:complexType name="uncertain-type">
    <xsd:attribute name="value" type="xsd:string" use="required">
    </xsd:attribute>
    </xsd:complexType>
    More(24%)
    Please help to resolve this problem, any suggestion is appreciate.

    We have confirmed this as a limitation in the current version 10.2.0 of the software, and earlier versions. We will provide more information on the timing and availability of a fix in the future.

  • Can SharePoint 2010 read data from an external SQL Server 2012 data Source?

    Hi,
     I would like to Know whether SharePoint can read the data from a SQL server 2012 external data source?
    I am not talking about SharePoint internal database. I need to get the data from an sql database which is used by some other application. For My SharePoint server I am using SQL 2008 R2 as an internal database engine. But I need to get some other data from
    another data Source which is configured in SQL server 2012.  Can any one help me on this whether is there any problem on accessing the data from SQL 2012. If there is no problem for that, please provide me the version of SQL data source compatible with
    SharePoint 2010 and SP 2010 sp1. 
    Thanks!
    Regards,
    Henoy 

    Hi Romeo ,
    I have already visited this page. But there is no answer for my question. I just want to know whether we
    can done the same from a SQL 2012 server.
    Please help me to know whether SharePoint 2010 is Compatible to get the data from SQL 2012 external data source.
    Thanks for your instant reply.
    Regards,
    Henoy TM.
    +919035752610

  • Data source to be created on CDPOS table in CRM.

    Hi
    I have to create a data source on CDPOS table in CRM. I would like to know is it necessary
    for  me to include or join it in infoset query with CDHDR table.
    In my case I do not want to extract all the records from CDPOS , only the ones with the
    date fields changed and which has tab_key as *zdtcvn(say). The valid records matching this criteria
    won't be too much. So still do i need to take into consideration the CDHDR table to improve the
    index thing...
    Please advise.
    Thanks in advance
    Regards,
    kate.

    Hi Kate,
      This will be depend upon your requirement , if u need to extract header level info  also like who changed this document what is the time...e.t.c..THEN u need to add CDHDR also..Otherwise CDPOS is enough....
      Assign points , If useful...

  • Shared Data Source with prompted credentials - errors out.

    Using Data Tools (VS 2010) to build report.
    Have built a simple report while trying to diagnose this issue.  This report is created in a project that was originally built in BIDS2008 and migrated to Data Tools 2010.  This project contains an original shared dataset that does NOT
    error out.
    Report contains
         1) shared data source with properties set to prompted credentials.
         2) shared dataset using data source above
         3) 2 parameters (month and year)
         4) Report uses shared dataset
         5) 2 text boxes in the report body to display the parameters values.
    When I try to preview this report I get the error:
         The execution failed for the shared data set "xxx"
         Cannot create a connection to data source 'Data source for shared dataset'
         Security processing failed with reason "3" ("password missing")
    Now, if I changed that same shared dataset to use a static login and password; the report renders fine.
    What is going on???

    Hi RoseNeedsAVacation,
    When you use shared data source with "Prompt for credential" option, please use SQL Server login credential to u view the report in BIDS or SSDT report designer environment.
    Note: The shared data source will automatic switch to SQL Server authentication mode if we use "Prompt for credential" option.
    After you deploy your report and shared data source to report server, please remember to configure your shared data source to use "Credentials supplied by the user running the report" option. Furthermore, select "Use
    as Windows credentials when connecting to the data source" checkbox if the credentials that the user provides are Windows Authentication credentials. Do not select this check box if you are using database authentication (for example, SQL
    Server Authentication).
    For more information, please refer to the article below:
    New Data Source Page (Report Manager):
    http://technet.microsoft.com/en-us/library/ms180077(v=sql.100).aspx
    Regards,
    Elvis Long
    TechNet Community Support

  • Can not see new Power Query data sources?

    Hi, 
    New Power Query can connect to three new data sources: Sybase IQ, Exchange, Dynamics CRM Online
    http://office.microsoft.com/en-us/excel-help/whats-new-in-power-bi-ha104149776.aspx
    I just installed the latest Power Query and wanted to connect to CRM Online instance. For some reason I can not see these new data sources, what am I missing?
    Thanks in advance!

    Hmm... I didn't think we had quite released this version yet. It's possible that that website went live before the link to the new download. There should be a blog entry about this on the "Power BI" blog in the very near future.
    At the risk of providing a spoiler :), the CRM Online support is via OData.

  • Error creating a new Data Source for Oracle connection

    I'm trying to create a new data source connection to our
    Oracle 10gR2 server using ColdFusion Administrator 6.1 Standard on
    a Windows 2003 server.
    I created a System DSN ODBC connection for Oracle. I can use
    the ODBC connection when connecting an Access db to Oracle but when
    I try to use it w/ ColdFusion Admin it generates the following
    error message:
    Connection verification failed for data source: sitest2
    []java.sql.SQLException: [Macromedia][SequeLink JDBC
    Driver][ODBC Socket]internal error:
    Driver's SQLAllocHandle on SQL_HANDLE_ENV failed
    The root cause was that: java.sql.SQLException:
    [Macromedia][SequeLink JDBC Driver][ODBC Socket]internal error:
    Driver's SQLAllocHandle on SQL_HANDLE_ENV failed
    Any help or recommendations would be greatly appreciated.
    Tina

    The schema/user is system and the password is alpha-numeric, upper/lower case, no special char ...
    As a workaround I was able to add the password to the entry in the datasources.xml file and restarted BI Pub.
    The JDBC is working ... I can test the connection successfully ... just wont save the entry
    Edited by: tread on Oct 17, 2011 11:55 AM

  • How to find all uses of data source views in an SSIS solution

    I am upgrading Visual Studio from 2008 to 2013 (with SSDT) and SQL Server 2008 R2 to  2012.  I have a solution with over 30 dstx files. Each file has multiple OLE data sources and lookup tasks.  There is inconsistent usage of data source views
    throughout (as compared to SQL queries or table references).  It is my understanding that all data source views need to be removed before upgrading SSIS packages from BIDS to SSDT.  I tried searching the files as XML for the DSVs but it appears the
    GUID reference changes per dstx file.  It seems like I will have to look at each source/lookup. Is there quicker way to search for where they are used? 
    Thanks
    Adding this question here.  Posted question incorrectly in VSO forum.

    All right, yes, they are dropped
    Never upgraded a package that had them.
    What happens if you just upgrade leaving a copy?
    Arthur My Blog
    This ended up being what I did for many of the packages.  Upgrading the packages severed the data source views and left the SQL in the related tasks (e.g. OLE Source task).  Sorry for the delayed mark as answered.

Maybe you are looking for