SQL Azure indexer support for Collection(Edm.String)

Is there a plan to support "Collection(Edm.String)
" with a SQL Azure indexer? Maybe via an XML type?
Unless I'm misunderstanding the supported types
https://msdn.microsoft.com/en-us/library/azure/dn946880.aspx
It sort of fizzles out after "time, timespan" but I'm assuming it's Not Supported's all the way down

http://feedback.azure.com/forums/263029-azure-search/suggestions/7189214-sql-azure-indexer-support-for-collection-edm-strin
Was going to start there but just wanted to vet that it indeed wasn't there.... I realize it's a bit awkward and anti - sql storing data like that in a column and will probably annoy DBAs.
 Currently the data we'd use this for would be Customer phone numbers, addresses, VIN for vehicles, and some account numbers... So nothing super fancy...   Straight delimiters might get funky with addresses but maybe a standard backspace escape
sequence or letting user use ascii hexcode if the delimiter is in the text..  
For now we already have a comma separated SearchText field we've indexed for use with FTS and I just pointed an Edm.string at that column in our DB and it seems to pick up all the comma separated elements... But I'm guessing it's not as efficient as if it
was stored in more specific collections.

Similar Messages

  • Bad index type for collection access ???

    Hi there,
    I am writing a script in VBA to connect to SAP.
    I encountered something strange.
    Why does this work:
    Set oSession = oConnection.Children(0)
    But this isn't:
    X = 0
    Set oSession = oConnection.Children(X)
    It results in an error: "Bad index type for collection access"
    Regards, Bas Prins

    Thanks,
    Although that is not proper VBA syntax I understand your suggestion.
    In VBA that would be:
    DIM X AS INTEGER
    X = 0
    Set oSession = oConnection.Children(X)
    But I tried several datatypes, all resulted in error.
    Regards, Bas.

  • Error connecting SQL Azure - Network access for Distributed Transaction Manager (MSDTC) has been disabled

    Sometimes I have an error connecting SQL Azure. The error occurs in an asp.net application and in a windows service running on VM in Azure. Error details:
    System.Data.Entity.Core.EntityException: The underlying provider failed on Open. ---> System.Transactions.TransactionManagerCommunicationException: Network access for Distributed Transaction Manager (MSDTC) has been disabled. Please enable DTC for network
    access in the security configuration for MSDTC using the Component Services Administrative tool. ---> System.Runtime.InteropServices.COMException: The transaction manager has disabled its support for remote/network transactions. (Exception from HRESULT:
    0x8004D024)
       at System.Transactions.Oletx.IDtcProxyShimFactory.ReceiveTransaction(UInt32 propgationTokenSize, Byte[] propgationToken, IntPtr managedIdentifier, Guid& transactionIdentifier, OletxTransactionIsolationLevel& isolationLevel,
    ITransactionShim& transactionShim)
       at System.Transactions.TransactionInterop.GetOletxTransactionFromTransmitterPropigationToken(Byte[] propagationToken)
       --- End of inner exception stack trace ---
       at System.Transactions.Oletx.OletxTransactionManager.ProxyException(COMException comException)
       at System.Transactions.TransactionInterop.GetOletxTransactionFromTransmitterPropigationToken(Byte[] propagationToken)
       at System.Transactions.TransactionStatePSPEOperation.PSPEPromote(InternalTransaction tx)
       at System.Transactions.TransactionStateDelegatedBase.EnterState(InternalTransaction tx)
       at System.Transactions.EnlistableStates.Promote(InternalTransaction tx)
       at System.Transactions.Transaction.Promote()
       at System.Transactions.TransactionInterop.ConvertToOletxTransaction(Transaction transaction)
       at System.Transactions.TransactionInterop.GetExportCookie(Transaction transaction, Byte[] whereabouts)
       at System.Data.SqlClient.SqlInternalConnection.EnlistNonNull(Transaction tx)
       at System.Data.ProviderBase.DbConnectionPool.PrepareConnection(DbConnection owningObject, DbConnectionInternal obj, Transaction transaction)
       at System.Data.ProviderBase.DbConnectionPool.TryGetConnection(DbConnection owningObject, UInt32 waitForMultipleObjectsTimeout, Boolean allowCreate, Boolean onlyOneCheckConnection, DbConnectionOptions userOptions, DbConnectionInternal&
    connection)
       at System.Data.ProviderBase.DbConnectionPool.TryGetConnection(DbConnection owningObject, TaskCompletionSource`1 retry, DbConnectionOptions userOptions, DbConnectionInternal& connection)
       at System.Data.ProviderBase.DbConnectionFactory.TryGetConnection(DbConnection owningConnection, TaskCompletionSource`1 retry, DbConnectionOptions userOptions, DbConnectionInternal oldConnection, DbConnectionInternal& connection)
       at System.Data.ProviderBase.DbConnectionInternal.TryOpenConnectionInternal(DbConnection outerConnection, DbConnectionFactory connectionFactory, TaskCompletionSource`1 retry, DbConnectionOptions userOptions)
       at System.Data.SqlClient.SqlConnection.TryOpenInner(TaskCompletionSource`1 retry)
       at System.Data.SqlClient.SqlConnection.TryOpen(TaskCompletionSource`1 retry)
       at System.Data.SqlClient.SqlConnection.Open()
       at System.Data.Entity.SqlServer.DefaultSqlExecutionStrategy.<>c__DisplayClass1.<Execute>b__0()
       at System.Data.Entity.SqlServer.DefaultSqlExecutionStrategy.Execute[TResult](Func`1 operation)
       at System.Data.Entity.Core.EntityClient.EntityConnection.Open()
       --- End of inner exception stack trace ---
       at System.Data.Entity.Core.EntityClient.EntityConnection.Open()
       at System.Data.Entity.Core.Objects.ObjectContext.EnsureConnection()
       at System.Data.Entity.Core.Objects.ObjectContext.ExecuteInTransaction[T](Func`1 func, IDbExecutionStrategy executionStrategy, Boolean startLocalTransaction, Boolean releaseConnectionOnSuccess)
       at System.Data.Entity.Core.Objects.ObjectQuery`1.<>c__DisplayClassb.<GetResults>b__9()
       at System.Data.Entity.SqlServer.DefaultSqlExecutionStrategy.Execute[TResult](Func`1 operation)
       at System.Data.Entity.Core.Objects.ObjectQuery`1.GetResults(Nullable`1 forMergeOption)
       at System.Data.Entity.Core.Objects.DataClasses.EntityReference`1.Load(MergeOption mergeOption)
       at System.Data.Entity.Core.Objects.DataClasses.RelatedEnd.DeferredLoad()
       at System.Data.Entity.Core.Objects.Internal.LazyLoadBehavior.LoadProperty[TItem](TItem propertyValue, String relationshipName, String targetRoleName, Boolean mustBeNull, Object wrapperObject)
       at System.Data.Entity.Core.Objects.Internal.LazyLoadBehavior.<>c__DisplayClass7`2.<GetInterceptorDelegate>b__2(TProxy proxy, TItem item)

    Hello,
    I am not an expert in MSDTC but as we know,SQL Azure Database does not support
    distributed transactions. This means that SQL Azure doesn’t allow Microsoft Distributed Transaction Coordinator (MS DTC) to delegate distributed transaction handling.
    One common cause of MSDTC getting involved in Entity Framework applications is the fact that we close and reopen the same connection as needed (i.e. for each query that is executed).To avoid the stack from opening and closing the connection multiple times,
    you can simply open the connection explicitly and run the queries in the same connectio.
    The following thread is about a similar issue, please refer to:
    http://answers.flyppdevportal.com/categories/azure/sqlazure.aspx?ID=d705a8cf-cba4-494c-96f6-96a136bd29e3
    What's more, you can also try the workaround that involves setting the Enlist option of the SQL Azure connection to false. For the detail explanation, please refer to:Entity
    FrameWork and SQL Azure
    Regards,
    Fanny Liu
    Fanny Liu
    TechNet Community Support

  • How is it possible to use Index Seek for LIKE %search-string% case?

    Hello,
    I have the following SP:
    CREATE PROCEDURE dbo.USP_SAMPLE_PROCEDURE(@Beginning nvarchar(15))
    AS
    SELECT * FROM HumanResources.Employee
    WHERE NationalIDNumber LIKE @Beginning + N'%';
    GO
    If I run the sp first time with param: N'94', then the following plan is generated and added to the cache:
    SQL Server "sniffs" the input value (94) when compiling the query. So for this param using Index Seek for AK_Employee_NationalIDNumber index will be the best option. On the other hand, the query plan should be generic enough to be able to handle
    any values specified in the @Beginning param.
    If I call the sp with @Beginning =N'%94':
    EXEC dbo.USP_SAMPLE_PROCEDURE N'%94'
    I see the same execution plan as above. The question is how is it possible to reuse this execution plan in this case? To be more precise, how
    Index Seek can be used in case LIKE %search-string% case. I expected that
    ONLY Index Scan operation can be used here.
    Alexey

    The key is that the index seek operator includes both seek (greater than and less than) and a predicate (LIKE).  With the leading wildcard, the seek is effectively returning all rows just like a scan and the filter returns only rows matching
    the LIKE expression.
    Do you want to say that in case of leading wildcard, expressions Expr1007 and Expr1008 (see image below) calculated such a way that
    Seek Predicates retrieve all rows from the index. And only
    Predicate does the real job by taking only rows matching the Like expression? If this is the case, then it explains how
    Index Seek can be used to resolve such queries: LIKE N'%94'.
    However, it leads me to another question: Since
    Index Seek in
    this particular case scans
    all the rows, what is the difference between
    Index Seek and Index Scan?
    According to
    MSDN:
    The Index Seek operator uses the seeking ability of indexes to retrieve rows from a nonclustered index.
    The storage engine uses the index to process
    only those rows that satisfy the SEEK:() predicate. It optionally may include a WHERE:() predicate, which the storage engine will evaluate against all rows that satisfy the SEEK:() predicate (it does not use the indexes to do this).
    The Index Scan operator retrieves
    all rows from the nonclustered index specified in the Argument column. If an optional WHERE:() predicate appears in the Argument column, only those rows that satisfy the predicate are returned.
    It seems like Index Scan is a special case of Index Seek,
    which means that when we see Index Seek in the execution plan, it does NOT mean that storage engine does NOT scan all rows. Right?
    Alexey

  • SQL Server 2014 supported for scheduled refresh?

    The following link, which lists the supported data sources for scheduled refresh in Power BI, lists all versions of SQL Server from 2005 to 2012:
    https://support.office.com/en-gb/article/Supported-Data-Sources-and-Data-Types-cb69a30a-2225-451f-a9d0-59d24419782e?ui=en-US&rs=en-GB&ad=GB
    However 2014 is not listed. Is this an oversight or is SQL Server 2014 really not supported?
    Thanks,
    Chris
    Check out my MS BI blog I also do
    SSAS, PowerPivot, MDX and DAX consultancy and run
    public SQL Server and BI training courses in the UK

    Chris, have you heard about this?
    I'm checking with our tech writing team. 
    Thanks!
    Ed Price, Azure & Power BI Customer Program Manager (Blog,
    Small Basic,
    Wiki Ninjas,
    Wiki)
    Answer an interesting question?
    Create a wiki article about it!

  • Web Services support for collection like ArrayList

    Getting this error message:
    "<Info> Did not find property empty in java.util.ArrayList:
    java.lang.NullPointerException"
    Are collections like ArrayList supported or am I trying to do something that
    can't be done? Any suggestions?
    Thanks
    Danny

    Hi Danny,
    Betcha thought I forgot 'bout cha, huh?
    I didn't ;-)
    The short answer to your question is no; WLS 6.1 does not support using any of
    the classes in java.util.*. However, it does support Object arrays as an element
    of an Object array, which means you can have code in you service implementation
    that looks like this:
    public Object[] getNestedObjects()
         Object[] objects = new Object[2];
         objects[0] = new Object[]{
              new String("23.76"),
              new Float((float)23.76)
         // Here, WorkOrder is a JavaBean or an object
         // that implements weblogic.soap.xml.XMLizable
         WorkOrder[] workOrders = new WorkOrder[]{
                        new WorkOrder(),
                        new WorkOrder(),
                        new WorkOrder()
         objects[1] = new Object[]{
              new String("My name is"),
              new String("Slim Shady"),
              workOrders
         return objects;
    The main issue I see with doing this is again, interoperability ;-)
    If the SOAP toolkit consuming the WSDL for the web service with the above method
    cannot handle the "xsd:anyType" XML Schema data type, you're in trouble. In general,
    SOAP toolkits want to know what type an element is so they can process it correctly.
    When you use a Variant data type (in Visual Basic) or java.lang.Object (in Java),
    you are basically promoting the use of weak data typing (or late binding). From
    what I've read about SOAP (well, actually what Don Box said he had in mind), XML
    Schema and namespaces were adopted to combat this very thing- weak or no data
    typing. Personally, I agree ;-) I think strong data typing makes processing a
    SOAP request much easier and faster, for everyone. Don't get me wrong, I don't
    have anything against object casting, but SOAP parameters (and return values)
    are not objects. They are XML elements, just like the other parts of the SOAP
    message. There is no behavior, just state. This being the case, isn't it safer
    to stick with the data types (e.g. arrays, structures, primitives, etc.) defined
    in the SOAP spec and "XML Schema Part 2: Datatypes" document? Enough SOAP boxing
    (pun intended). If you are interested in the "complexTypes" object graphs (using
    developer defined classes that implement weblogic.soap.xml.XMLizable) lab results,
    I can send you a zip. The biggest difference between using JavaBeans and objects
    that implement weblogic.soap.xml.XMLizable, is that you have much more control
    over which Java objects the later gets serialized/deserialized to/from. With classes
    that implement weblogic.soap.xml.XMLizable, the SOAP processor passes you the
    actual XML stream for the SOAP parameter (which in this case, is your developer
    defined class). It (the SOAP processor) also calls your developer defined class
    to get an XML representation for it's data types. It works great with Java clients
    (using the client.jar, of course), and I'm trying it out with .NET and MS SOAP
    clients this weekend. The most interesting revelation was figuring out how wsgen
    (well, actually the WSDL processor) generates the information in the <types> element
    of the WSDL ;-) I actually even know how to code an XMLizable so that it produces
    the exact element and attribute names I want. I also figure out how to get it
    to produce an <element ...>, as opposed to a <attribute ...>, and visa versa.
    Of course, WLS 7.0 (and WebLogic Workshop) make all of this a moot point, but
    it was fun to get the SOAP/WSDL processors in WLS 6.1 to "do my bidding" for a
    short while ;-)
    Regards,
    Mike Wooten
    "Danny Ngo" <[email protected]> wrote:
    >
    Hi Mike,
    Thank you for your response. Do you know if WLS 6.1 Web Services support
    the
    return of nested object array (object that contains other object arrays).
    Thanks
    -Danny
    "Michael Wooten" <[email protected]> wrote:
    Hi Danny,
    For interoperability (with non-Java SOAP implementations)reasons, the
    Java collections
    are not supported in WLS 6.1 ;-) You should consider using an arrayin
    place of
    this.
    Regards,
    Mike Wooten
    Danny Ngo <[email protected]> wrote:
    Getting this error message:
    "<Info> Did not find property empty in java.util.ArrayList:
    java.lang.NullPointerException"
    Are collections like ArrayList supported or am I trying to do something
    that
    can't be done? Any suggestions?
    Thanks
    Danny

  • Is Windows Azure officially supported for Exchange 2013?

    Though Exchange 2013 can be installed on Windows Azure, however can somebody help me in knowing that if its officially supported with it OR can live environment of Exchange be setup on Windows Azure?

    Hinte is correct.
    You can set up a Exchange 2013 lab in Azure. You may have seen that Henrik Walther has done just that:
    http://www.msexchange.org/articles-tutorials/office-365/exchange-online/deploying-exchange-2013-hybrid-lab-environment-windows-azure-part1.html
    But this is just for practice / training.
    Besides lack of support, cost of sufficient resources to run E2K13 correctly would be greater than simply using O365 / Exchange Online.
    Please mark as helpful if you find my contribution useful or as an answer if it does answer your question. That will encourage me - and others - to take time out to help you.

  • SQL Server 2014 support for Workflow Manager?

    Is this configuration supported?
    The prerequisite requirements for Workflow Manager are out of date and do not even list SQL 2012 R2 as a supported DB server.

    Hi JB
    We installed Workflow Manager 1.0 on a Dev SharePoint 2013 SP1 + April 2014 CU Farm running Against SQLServer 2014 which I suspect in not much different under the hood to your SQL2012R2.
    Freelance consultant

  • SQL Server Enterprise support for Offline / Online Application development

    Hi,
    We have suggested to use the SQL Server 2012 as Database for our Enterprise application. We have the Offline/Online two set of application. Can you suggest any offline database to sync the Online SQL Server database.
    Thanks,
    Sethu

    What is an offline database?
    One can setup an "online" database and an "offline" database using SQL Server 2012 Enterprise Edition.
    SSIS packages can be used to synch data between the databases:
    http://blogs.msdn.com/b/jorgepc/archive/2010/12/07/synchronize-two-tables-using-sql-server-integration-services-ssis-part-i-of-ii.aspx?Redirected=true
    Kalman Toth Database & OLAP Architect
    SELECT Video Tutorials 4 Hours
    New Book / Kindle: Exam 70-461 Bootcamp: Querying Microsoft SQL Server 2012

  • Connection between two SQL Azure Databases

    We have a requirement to move data (partial data in a table based on policy conditions) between two SQL Azure Databases. Want to know the best possible way to do this.
    We are not looking at Data Sync Framework - as this is only a Preview version and we have to use this in an ongoing basis in Production and the volume of data is quite high.
    The option that we have currently is to use an on premise stored procedure - that will have two link servers to the source and target SQL Azure Databases and do the data movement in one transaction.
    Are there any other better options to do this ? Any pointers will be helpful.

    Hi Kothai Ramanathan,
    According to your description, if you just want to moving part of the data from a huge table between two different SQL Server database, you can use
    SQL Server replication to sync the part of data via articles. However, in SQL Azure database, it does not support SQL Server replication, if you want to migrate database, as other post, you
    can use Data-Tier Application Import and Export or other ways.
     In addition, for just moving the part of data,  you can also create two Linked Servers between local SQL Server database and two different SQL azure database. For example, you can get the changed data from the first Linked
    Server and saved the data to Local database, then insert these data to the other azure database via the second Linked Server.
     For more information, see:
    http://azure.microsoft.com/blog/2012/09/19/announcing-updates-to-windows-azure-sql-database/
    http://blogs.msdn.com/b/sqlcat/archive/2011/03/08/linked-servers-to-sql-azure.aspx
    Regards,
    Sofiya Li
    Sofiya Li
    TechNet Community Support

  • Best way to Insert Millions records in SQL Azure on daily basis?

    I am maintaining millions of records in Sql Server 2008 R2 and now i am intended to migrate these on SQL Azure.
    In existing system with SQL Server 2008 R2, few SSIS packages and Stored Procedures are firstly truncate the existing records and then perform Insert operation on the table which holds
    approx 26 Million records in 30 mins. on Daily basis (as system demands).
    When i migrate these on SQL Azure, i am unable to perform these operations in a
    faster way as i did in SQL 2008. Sometimes i got Request timeout error.
    While searching for faster way, many of them suggest for Batch process or BCP. But Batch processing is NOT suitable in my case because it takes much time to insert those records. I required some faster and efficient way on SQL Azure.
    Hoping for some good suggestions.
    Thanks in advance :)
    Ashish Narnoli

    +1 to Frank's advice.
    Also, please upgrade your Azure SQL Database server to
    V12 as you will receive higher performance on the premium tiers.  As you scale-up your database for your bulk insert, remember that
    SQL Database charges by the hour. To minimize costs, scale back down when the inserts have completed.

  • TREX support for IDOCs only no PI involved

    I am aware of the TREX index support for PI XML content.
    What if only IDOCs are used. No PI.
    We09 is timing out. So we need a better index.
    Is Trex and option ?

    Hi Che,
    Please have a look at below blog
    Principal Propagation using SAP Assertion Ticket CRM -> PO7.31 Single Stack
    regards,
    Harish

  • Selective XML Index feature is not supported for the current database version , SQL Server Extended Events , Optimizing Reading from XML column datatype

    Team , Thanks for looking into this  ..
    As a last resort on  optimizing my stored procedure ( Below ) i wanted to create a Selective XML index  ( Normal XML indexes doesn't seem to be improving performance as needed ) but i keep getting this error within my stored proc . Selective XML
    Index feature is not supported for the current database version.. How ever
    EXECUTE sys.sp_db_selective_xml_index; return 1 , stating Selective XML Indexes are enabled on my current database .
    Is there ANY alternative way i can optimize below stored proc ?
    Thanks in advance for your response(s) !
    /****** Object: StoredProcedure [dbo].[MN_Process_DDLSchema_Changes] Script Date: 3/11/2015 3:10:42 PM ******/
    SET ANSI_NULLS ON
    GO
    SET QUOTED_IDENTIFIER ON
    GO
    -- EXEC [dbo].[MN_Process_DDLSchema_Changes]
    ALTER PROCEDURE [dbo].[MN_Process_DDLSchema_Changes]
    AS
    BEGIN
    SET NOCOUNT ON --Does'nt have impact ( May be this wont on SQL Server Extended events session's being created on Server(s) , DB's )
    SET TRANSACTION ISOLATION LEVEL READ UNCOMMITTED
    select getdate() as getdate_0
    DECLARE @XML XML , @Prev_Insertion_time DATETIME
    -- Staging Previous Load time for filtering purpose ( Performance optimize while on insert )
    SET @Prev_Insertion_time = (SELECT MAX(EE_Time_Stamp) FROM dbo.MN_DDLSchema_Changes_log ) -- Perf Optimize
    -- PRINT '1'
    CREATE TABLE #Temp
    EventName VARCHAR(100),
    Time_Stamp_EE DATETIME,
    ObjectName VARCHAR(100),
    ObjectType VARCHAR(100),
    DbName VARCHAR(100),
    ddl_Phase VARCHAR(50),
    ClientAppName VARCHAR(2000),
    ClientHostName VARCHAR(100),
    server_instance_name VARCHAR(100),
    ServerPrincipalName VARCHAR(100),
    nt_username varchar(100),
    SqlText NVARCHAR(MAX)
    CREATE TABLE #XML_Hold
    ID INT NOT NULL IDENTITY(1,1) PRIMARY KEY , -- PK necessity for Indexing on XML Col
    BufferXml XML
    select getdate() as getdate_01
    INSERT INTO #XML_Hold (BufferXml)
    SELECT
    CAST(target_data AS XML) AS BufferXml -- Buffer Storage from SQL Extended Event(s) , Looks like there is a limitation with xml size ?? Need to re-search .
    FROM sys.dm_xe_session_targets xet
    INNER JOIN sys.dm_xe_sessions xes
    ON xes.address = xet.event_session_address
    WHERE xes.name = 'Capture DDL Schema Changes' --Ryelugu : 03/05/2015 Session being created withing SQL Server Extended Events
    --RETURN
    --SELECT * FROM #XML_Hold
    select getdate() as getdate_1
    -- 03/10/2015 RYelugu : Error while creating XML Index : Selective XML Index feature is not supported for the current database version
    CREATE SELECTIVE XML INDEX SXI_TimeStamp ON #XML_Hold(BufferXml)
    FOR
    PathTimeStamp ='/RingBufferTarget/event/timestamp' AS XQUERY 'node()'
    --RETURN
    --CREATE PRIMARY XML INDEX [IX_XML_Hold] ON #XML_Hold(BufferXml) -- Ryelugu 03/09/2015 - Primary Index
    --SELECT GETDATE() AS GETDATE_2
    -- RYelugu 03/10/2015 -Creating secondary XML index doesnt make significant improvement at Query Optimizer , Instead creation takes more time , Only primary should be good here
    --CREATE XML INDEX [IX_XML_Hold_values] ON #XML_Hold(BufferXml) -- Ryelugu 03/09/2015 - Primary Index , --There should exists a Primary for a secondary creation
    --USING XML INDEX [IX_XML_Hold]
    ---- FOR VALUE
    -- --FOR PROPERTY
    -- FOR PATH
    --SELECT GETDATE() AS GETDATE_3
    --PRINT '2'
    -- RETURN
    SELECT GETDATE() GETDATE_3
    INSERT INTO #Temp
    EventName ,
    Time_Stamp_EE ,
    ObjectName ,
    ObjectType,
    DbName ,
    ddl_Phase ,
    ClientAppName ,
    ClientHostName,
    server_instance_name,
    nt_username,
    ServerPrincipalName ,
    SqlText
    SELECT
    p.q.value('@name[1]','varchar(100)') AS eventname,
    p.q.value('@timestamp[1]','datetime') AS timestampvalue,
    p.q.value('(./data[@name="object_name"]/value)[1]','varchar(100)') AS objectname,
    p.q.value('(./data[@name="object_type"]/text)[1]','varchar(100)') AS ObjectType,
    p.q.value('(./action[@name="database_name"]/value)[1]','varchar(100)') AS databasename,
    p.q.value('(./data[@name="ddl_phase"]/text)[1]','varchar(100)') AS ddl_phase,
    p.q.value('(./action[@name="client_app_name"]/value)[1]','varchar(100)') AS clientappname,
    p.q.value('(./action[@name="client_hostname"]/value)[1]','varchar(100)') AS clienthostname,
    p.q.value('(./action[@name="server_instance_name"]/value)[1]','varchar(100)') AS server_instance_name,
    p.q.value('(./action[@name="nt_username"]/value)[1]','varchar(100)') AS nt_username,
    p.q.value('(./action[@name="server_principal_name"]/value)[1]','varchar(100)') AS serverprincipalname,
    p.q.value('(./action[@name="sql_text"]/value)[1]','Nvarchar(max)') AS sqltext
    FROM #XML_Hold
    CROSS APPLY BufferXml.nodes('/RingBufferTarget/event')p(q)
    WHERE -- Ryelugu 03/05/2015 - Perf Optimize - Filtering the Buffered XML so as not to lookup at previoulsy loaded records into stage table
    p.q.value('@timestamp[1]','datetime') >= ISNULL(@Prev_Insertion_time ,p.q.value('@timestamp[1]','datetime'))
    AND p.q.value('(./data[@name="ddl_phase"]/text)[1]','varchar(100)') ='Commit' --Ryelugu 03/06/2015 - Every Event records a begin version and a commit version into Buffer ( XML ) we need the committed version
    AND p.q.value('(./data[@name="object_type"]/text)[1]','varchar(100)') <> 'STATISTICS' --Ryelugu 03/06/2015 - May be SQL Server Internally Creates Statistics for #Temp tables , we do not want Creation of STATISTICS Statement to be logged
    AND p.q.value('(./data[@name="object_name"]/value)[1]','varchar(100)') NOT LIKE '%#%' -- Any stored proc which creates a temp table within it Extended Event does capture this creation statement SQL as well , we dont need it though
    AND p.q.value('(./action[@name="client_app_name"]/value)[1]','varchar(100)') <> 'Replication Monitor' --Ryelugu : 03/09/2015 We do not want any records being caprutred by Replication Monitor ??
    SELECT GETDATE() GETDATE_4
    -- SELECT * FROM #TEMP
    -- SELECT COUNT(*) FROM #TEMP
    -- SELECT GETDATE()
    -- RETURN
    -- PRINT '3'
    --RETURN
    INSERT INTO [dbo].[MN_DDLSchema_Changes_log]
    [UserName]
    ,[DbName]
    ,[ObjectName]
    ,[client_app_name]
    ,[ClientHostName]
    ,[ServerName]
    ,[SQL_TEXT]
    ,[EE_Time_Stamp]
    ,[Event_Name]
    SELECT
    CASE WHEN T.nt_username IS NULL OR LEN(T.nt_username) = 0 THEN t.ServerPrincipalName
    ELSE T.nt_username
    END
    ,T.DbName
    ,T.objectname
    ,T.clientappname
    ,t.ClientHostName
    ,T.server_instance_name
    ,T.sqltext
    ,T.Time_Stamp_EE
    ,T.eventname
    FROM
    #TEMP T
    /** -- RYelugu 03/06/2015 - Filters are now being applied directly while retrieving records from BUFFER or on XML
    -- Ryelugu 03/15/2015 - More filters are likely to be added on further testing
    WHERE ddl_Phase ='Commit'
    AND ObjectType <> 'STATISTICS' --Ryelugu 03/06/2015 - May be SQL Server Internally Creates Statistics for #Temp tables , we do not want Creation of STATISTICS Statement to be logged
    AND ObjectName NOT LIKE '%#%' -- Any stored proc which creates a temp table within it Extended Event does capture this creation statement SQL as well , we dont need it though
    AND T.Time_Stamp_EE >= @Prev_Insertion_time --Ryelugu 03/05/2015 - Performance Optimize
    AND NOT EXISTS ( SELECT 1 FROM [dbo].[MN_DDLSchema_Changes_log] MN
    WHERE MN.[ServerName] = T.server_instance_name -- Ryelugu Server Name needes to be added on to to xml ( Events in session )
    AND MN.[DbName] = T.DbName
    AND MN.[Event_Name] = T.EventName
    AND MN.[ObjectName]= T.ObjectName
    AND MN.[EE_Time_Stamp] = T.Time_Stamp_EE
    AND MN.[SQL_TEXT] =T.SqlText -- Ryelugu 03/05/2015 This is a comparision Metric as well , But needs to decide on
    -- Peformance Factor here , Will take advise from Lance if comparision on varchar(max) is a vital idea
    --SELECT GETDATE()
    --PRINT '4'
    --RETURN
    SELECT
    top 100
    [EE_Time_Stamp]
    ,[ServerName]
    ,[DbName]
    ,[Event_Name]
    ,[ObjectName]
    ,[UserName]
    ,[SQL_TEXT]
    ,[client_app_name]
    ,[Created_Date]
    ,[ClientHostName]
    FROM
    [dbo].[MN_DDLSchema_Changes_log]
    ORDER BY [EE_Time_Stamp] desc
    -- select getdate()
    -- ** DELETE EVENTS after logging into Physical table
    -- NEED TO Identify if this @XML can be updated into physical system table such that previously loaded events are left untoched
    -- SET @XML.modify('delete /event/class/.[@timestamp="2015-03-06T13:01:19.020Z"]')
    -- SELECT @XML
    SELECT GETDATE() GETDATE_5
    END
    GO
    Rajkumar Yelugu

    @@Version : ----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------
    Microsoft SQL Server 2012 - 11.0.5058.0 (X64)
        May 14 2014 18:34:29
        Copyright (c) Microsoft Corporation
        Developer Edition (64-bit) on Windows NT 6.2 <X64> (Build 9200: ) (Hypervisor)
    (1 row(s) affected)
    Compatibility level is set to 110 .
    One of the limitation states - XML columns with a depth of more than 128 nested nodes
    How do i verify this ? Thanks .
    Rajkumar Yelugu

  • Azure indexer for dynamic packaged asset

    Hi I have an mp4 uploaded in the media services and I ran the Azure encoder to create a dynamic package of the video. I also ran the Azure indexer to create the index files for captions.
    I am using the Azure Media Player using the iframe code below in my webpages, however I do not see the captions. 
    <iframe src="//aka.ms/azuremediaplayeriframe?url={mp4 url}" height="300" width="200" frameborder="0" allowfullscreen/>
    I see the ttmland vtt files got created under the asset folder the source mp4 resided but nothing on the dynamically packaged asset. What is the correct way to reference the manifest to get the captions showing?
    Thanks for any help.

    Hi,
    If you deploy the application to local Tomcat, can you access the site using root URL? The StackOverflow thread states it is needed to rename the war file to ROOT.war, and remove the out-of-the-box ROOT/ directory. Please double check whether you have done
    that. You may also want to refer to
    http://tomcat.apache.org/tomcat-7.0-doc/config/context.html, where it states:
    The context path of this web application, which is matched against the beginning of each request URI to select the appropriate web application for processing. All of the context paths within a particular Host must be unique. If you specify a context path of
    an empty string (""), you are defining the default web application for this Host, which will process all requests not assigned to other Contexts.
    This attribute must only be used when statically defining a Context in server.xml. In all other circumstances, the path will be inferred from the filenames used for either the .xml context file or the docBase.
    Even when statically defining a Context in server.xml, this attribute must not be set unless either the docBase is not located under the Host's appBase or both deployOnStartup and autoDeploy are false. If this rule is not followed, double deployment is likely
    to result.
    And this link(http://stackoverflow.com/questions/16160565/configure-the-path-localhost-of-the-war-application-to-be-the-root-java-ee
    ) may help you.
    Thanks.
    Will
    <THE CONTENT IS PROVIDED "AS IS" WITHOUT WARRANTY OF ANY KIND, WHETHER EXPRESS OR IMPLIED>
    Thanks
    MSDN Community Support
    Please remember to "Mark as Answer" the responses that resolved your issue. It is a common way to recognize those who have helped you, and makes it easier for other visitors to find the resolution later.

  • Error when accessing the web-based portal for SQL Azure

    Hi,
    when trying to access SQL azure management portal, the following error is thrown:
    [MoreThanOneMatch]
    Arguments:
    Debugging resource strings are unavailable. Often the key and arguments provide sufficient information to diagnose the problem. See
    http://go.microsoft.com/fwlink/?linkid=106663&Version=5.1.30214.00&File=System.Core.dll&Key=MoreThanOneMatch -
    Looking at the site in question, the problem seems to be Silverlight related. However, I am unable to debug as suggested since the server is located on Microsoft premises. To rule out a wrong password, I did a reset and ruled out typos by copying username
    and password into the field. Furthermore since its not a production database, all IPv4 addresses (range 0.0.0.0-255.255.255.255) are allowed to access the database, so the problem should not be related to ip-based access restrictions. Is this a bug
    in the web interface or is the problem sitting in front of the screen?
    Regards,
    Chris

    Hello,
    I also meet this issue when connect to SQL database without specify database on the login page. But the error disappear when I specify the database name in the database section, for example, master.It may be a bug in Management Portal of SQL Database. 
    Regards,
    Fanny Liu
    If you have any feedback on our support, please click here. 
    Fanny Liu
    TechNet Community Support

Maybe you are looking for

  • Urgent help please.  Inner Join caused ora-00933 error

    I ran this one , works fine: SELECT DISTINCT EXP.EXP_ID, EXP.DATU_EXP_WIRE_CENTER_CLLI, EXP.DATU_EXP_IP, EXP.DATU_EXP_CLLI, EXP.DATU_EXP_PORT, EXP.DATU_EXP_NAME, EXP.DATU_EXP_CITY, EXP.DATU_EXP_STATE, EXP.DATU_EXP_SW_VERSION, DECODE(LAST_ALARM.LAST_A

  • Unable to upload updated photo gallery to server

    I am trying to add a new album to my Photo Gallery. I publish it and can view the new album from my desktop but when I upload the files to the server I receive the following message and the Photo Gallery cannot be found on-line: "Can't change directo

  • LightBox is opening in a new window!

    I have gone though so many other old discussion on this subject and have tried everything i have read about but nothing seemes to be working If anybody could have at look at my code below it would be so helpful maybe i have missed some think I would

  • My songs only play for 6 seconds

    My imported music is only playing for six seconds then going to the next track. All of my purchased music works fine though.  The imported files also play in their entirety on Windows Media Player so I know it's not the files' fault. Any suggestions?

  • Why I can not download and update software from App store?

    When I want to install some software, only display"please wait",I cannot download and update anyone! Please,help me!!!!