Issue with OLEDB Source in BIML

Hi all,
I am trying to automate a data flow with BIML. I am using an expression to build my SQL dynamically based on input parameter. Sparing
details of the use case, I need this flexibility in my project. I am having 1 master table which consists of file name and source query.
I want to use SQL Command from Variable Data access mode in OLEDB Source.
On the basis of input file source query need to change automatically .However; my query is not being evaluated when the package
is generated. The query will populate after package generation, when I open the source and set access mode SQL Command, but I cannot seem to get this configured automatically as desired. This is preventing me from doing transformations in the script.
Please help me to achieve the above scenario
Below is my BIML Script(C#).
<Biml
xmlns="http://schemas.varigence.com/biml.xsd">
<!-- Database Connection manager-->
<Connections>
<Connection
Name="Archive"
ConnectionString="Data Source=RLDEVOLP03.DEVELOPMENT.LOCAL;Initial Catalog=Archive;Provider=SQLNCLI11.1;Integrated Security=SSPI;Auto Translate=False;"
/>
<Connection
Name="DataStaging"
ConnectionString="Data Source=RLDEVOLP03.DEVELOPMENT.LOCAL;Initial Catalog=DataStaging;Provider=SQLNCLI11.1;Integrated Security=SSPI;Auto Translate=False;"
/>
</Connections>
<!-- Name Of the the Package-->
<Packages>
<Package
Name="LoadArchive Using BIML"
ConstraintMode="Linear"
ProtectionLevel="EncryptSensitiveWithUserKey" >
<Variables>
<Variable
Name="V_Archive_tablename"
DataType="String" ></Variable>
<Variable
Name="V_Archivequery"
DataType="String"
EvaluateAsExpression="true">SELECT a.*, b.BBxKey as Archive_BBxKey, b.RowChecksum as Archive_RowChecksum
FROM dbo.ImportBBxFbapp a LEFT OUTER JOIN Archive.dbo.ArchiveBBxFbapp b ON a.Col001 = b.BBxKey Where (b.LatestVersion = 1 OR b.LatestVersion IS NULL)
</Variable>
<Variable
Name="v_Src_FileName"
DataType="String" >FBAPP</Variable>
<!-- Load Data Truncate Staging Sequence Container--> 
<Container
Name="Load Data Truncate Staging"
ConstraintMode="Parallel">
<Tasks>
<Dataflow
Name="Archive Data"
DelayValidation="true" >
<Transformations>
<OleDbSource
Name="Source"
ConnectionName="DataStaging"
ValidateExternalMetadata="false">
<TableFromVariableInput
VariableName="User.V_Archivequery"/>
</OleDbSource>            
</Transformations>
</Dataflow>
</Tasks>
</Container>
</Tasks>
</Package>
</Packages>
</Biml>
Regards,
Vipin jha
Thankx & regards, Vipin jha MCP

Hi Vipin,
very few people would respond here, why don't you post to the BIML dedicated forum: http://varigence.com/Forums?forumName=Biml&threadID=0
Arthur
MyBlog
Twitter

Similar Messages

  • Latest Update of Adobe Media Encoder Issue with C4D Source

    Updated Media Encoder to Build 8.0.0.173 this AM when the Adobe notification came through and it now refuses to encode any project with C4D source.   The issue is codec agnostic.   Tried with both C4D R15 and R16.  No joy.  Any clue?  Would like to roll back to the previous version but not sure that option is available within the uber-slick CC Adobe world.

    I'm confused. You say that you updated ths morning, but then you give the version number for the previous version (8.0), rather than the current version for AME (8.1).
    I just tried rendering and exporting an After Effects composition containing a .c4d scene file through AME 8.1, and it works fine for me on Windows and Mac OS.

  • Issue with data source after deploying

    We are experiencing an issue with our data source after deployment of a cube. On the datasource properties in Visual Studio 2012, we have the max connections set to 0 before the deployment. Once the cube is deployed, I can navigate to the <name>.0.ds.xml
    file and open it and see that the <MaxActiveConnections>0</MaxActiveConnections> is indeed set to 0. At some point over the next couple days, a process of the cube or some other action causes that value to get updated to some number too large to
    be converted to an int, and makes the datasource invalid. At that point we cannot view the datasource properties in SSMS, we cannot open the cube project from Visual Studio, and we’ve even had failures when trying to process the cube.  Is there a config
    somewhere that would cause this value to get overwritten, or some other behind the scenes process that we can look at?
    Our server information is:
    Microsoft SQL Server 2012 (SP1) - 11.0.3153.0 (X64)
                    Jul 22 2014 15:26:36
                    Copyright (c) Microsoft Corporation
                    Enterprise Edition: Core-based Licensing (64-bit) on Windows NT 6.2 <X64> (Build 9200: ) (Hypervisor)
    Chad Dotzenrod SWC | TECHNOLOGY PARTNERS 1420 Kensington Road, Suite 110 Oak Brook, Illinois 60523-2144 http://www.swc.com

    Typically you would import the metadata from the source location and either use that location as the data source (and so not need to redeploy), or deploy it to a separate target location.
    The replace action is destructive as you've found, and effectively performs a drop table followed by create table. Hence any data in the table is lost.
    If you just want the Control Center Manager to correctly display that the table is deployed, try setting the action to "Upgrade". This will try to upgrade the deployed object to match the definition in OWB, but as the two are identical this will result in no changes. However, it will update the deployment records to indicate that the object is deployed.
    Nigel.

  • SSRS 2012 - Windows Server 2012 R2 issue with OLEDB Connection String

    Hoping someone can help. As stated above, we are running SQL Server 2012 with SP1 on Windows Server 2012 R2. When I create a Data Source as follows in Reporting Services, I receive the following error message. Any thoughts on why this fails? The same data
    source works on Windows Server 2008 R2 Enterprise SQL Server 2008 R2. 
    Data Source type: OLE DB
    Connection string: Provider=Microsoft.ACE.OLEDB.12.0;Data Source="\\networknamehere\CSVFiles";Mode=Read;Extended Properties="text;HDR=YES;FMT=CSVDelimited"
    ERROR: Throwing Microsoft.ReportingServices.ReportProcessing.ReportProcessingException: , Microsoft.ReportingServices.ReportProcessing.ReportProcessingException: Cannot create a connection to data source 'CSV_Datasource'. ---> System.Data.OleDb.OleDbException:
    Unspecified error
       at System.Data.OleDb.OleDbConnectionInternal..ctor(OleDbConnectionString constr, OleDbConnection connection)
       at System.Data.OleDb.OleDbConnectionFactory.CreateConnection(DbConnectionOptions options, Object poolGroupProviderInfo, DbConnectionPool pool, DbConnection owningObject)
       at System.Data.ProviderBase.DbConnectionFactory.CreateNonPooledConnection(DbConnection owningConnection, DbConnectionPoolGroup poolGroup)
       at System.Data.ProviderBase.DbConnectionFactory.GetConnection(DbConnection owningConnection)
       at System.Data.ProviderBase.DbConnectionClosed.OpenConnection(DbConnection outerConnection, DbConnectionFactory connectionFactory)
       at System.Data.OleDb.OleDbConnection.Open()
       at Microsoft.ReportingServices.DataExtensions.ConnectionExtension.Open()
       at Microsoft.ReportingServices.Diagnostics.DataExtensionConnectionBase.OpenConnection(IProcessingDataSource dataSourceObj, DataSourceInfo dataSourceInfo, IDbConnection conn)
       --- End of inner exception stack trace ---;

    Hi ABAA101,
    According to your description, when you use csv file as data source in SSRS 2012 report, you got the error message.
    I tested the issue in my local machine, due to some policy restrictions, I could not install Access in my local machine. To workaround the issue, we can use the connection string like below:
    Provider=Microsoft.Jet.OLEDB.4.0;Data Source=C:\\report;
    Extended Properties="text;HDR=Yes;FMT=Delimited"
    The csv file is stored in report folder in C driver.
    In Query text box in the dataset, we can use the query like below:
    Select * from new.csv
    If you have any more questions, please feel free to ask.
    Thanks,
    Wendy Fu
    Wendy Fu
    TechNet Community Support

  • Connectivity issue with Oracle Source

    Hi,
    Am trying to run a SSIS package to pull data from Oracle. Package was running fine from BIDS.But when i run the job from Sql Agent am getting the below error.Can you please help me on this.Am using 2012 version of SSIS
    Message
    Executed as user: ABCDEF\SQLDEV015. Microsoft (R) SQL Server Execute Package Utility  Version 11.0.5548.0 for 64-bit  Copyright (C) Microsoft Corporation. All rights reserved.    Started:  9:32:29 PM  Error: 2014-12-22 21:32:31.32
        Code: 0x000002BD     Source: Package4 Connection manager "ORACLE"     Description: Oracle Home not found.  End Error  Error: 2014-12-22 21:32:31.32     Code: 0x0000020F     Source: Connect
    to ORACLE Oracle Source [2]     Description: The AcquireConnection method call to the connection manager ORACLE failed with error code 0x80004005.  There may be error messages posted before this with more information on why the AcquireConnection
    method call failed.  End Error  Error: 2014-12-22 21:32:31.32     Code: 0xC0047017     Source: Connect to ORACLE SSIS.Pipeline     Description: Oracle Source failed validation and returned error code 0x80004005.  End
    Error  Error: 2014-12-22 21:32:31.32     Code: 0xC004700C     Source: Connect to ORACLE SSIS.Pipeline     Description: One or more component failed validation.  End Error  Error: 2014-12-22 21:32:31.32    
    Code: 0xC0024107     Source: Connect to ORACLE      Description: There were errors during task validation.  End Error  DTExec: The package execution returned DTSER_FAILURE (1).  Started:  9:32:29 PM  Finished:
    9:32:31 PM  Elapsed:  1.422 seconds.  The package execution failed.  The step failed.

    Hi SSDL,
    According to the error message, the issue is that the user cannot connect to the Oracle server. To fix this issue, please refer to the following suggestions:
    Verify the Oracle server is installed on the server that the connection string indicates. It means that verify the connection string is incorrect.
    Does the user that runs the package the correct
    authorization?
    Check if the SQL Agent job step that runs this package is set to run it in 32 bit mode.
    The following two similar threads are for your references:
    http://stackoverflow.com/questions/26790105/ssis-package-fails-in-sql-server-agent-but-not-in-visual-studio-odbc-connectio
    http://www.attunity.com/forums/microsoft-ssis-oracle-connector/issues-pulling-data-oracle-package-fails-1472.html
    Thanks,
    Katherine Xiong
    Katherine Xiong
    TechNet Community Support

  • Issues with XML Source Code

    Has anyone seen or heard fo this issue:
    I was sent a new xml source code and loaded it into my .rtf file. When I went to preview the file it showed up no problem. All great and dandy!
    When I tried to add a field and preview the field would not show up. Yes, there is data in the XML sourse code for the tag.
    I decided to run a test and load some of the old XML source code and had no problems. I was able to move fields, add plain text, and delete fields and my changes would display in the preview.
    The only thing we can think of that is different is that the previous xml source code was generated from 9. 2. 08 database we recently upgraded to 10g. The new source code is from 10g.
    Does this make any sense? Would our upgrade to 10g result in not being able to change the template? Or is it something else?
    Thanks,
    Jessica

    Tim,
    I actually just sent you an e-mail. It contains the rtf and the source codes.
    Thanks,
    Jessica

  • 10.1.3 Migration Issue with data-sources.xml

    Hi:
    Running into lots of migration issues for an application built in 10.1.3 EA, migrating to 10.1.3. production, mostly in the BC project. Here's one: the database connections built in EA were not carried over to prod, so they needed to get re-built. No biggie. But now the data-sources.xml file in the model project has duplicate entries for my data sources. I have tried directly editting that file and removing the duplicates, but when I go to run my application (ADF Faces & ADF BC) JDeveloper hangs. When I abort JDev and re-start it, the duplicate entries are back in data-sources.xml (and yes, I did save the file after making the changes). It seems that the xml file is built dynamically from definitions stored elsewhere. Where is the source? What will I need to modify to get the changes to take?
    Thanks.
    Johnny Lee

    There is new control over when/whether JDeveloper add/updates your data-sources.xml file for the embedded server. The settings are located here:
    Tools | Embedded OC4J Server Preferences | Current Workspace | Data Sources
    There are some settings there that control whether/when we update the data-sources.xml

  • Issue with Data sources

    Hi Gurus,
               I have issue like can I use one data source for more than one info source.
               And one more thing is it necessary to add key figures in Comm Struc, and in Tran. Struct.
               Please I am still in starting stage of BW. If any have idea give me a advice or email me to [email protected]
    Regds
    Dayakar Reddy.

    Hi SK,
          Acutally I need only two fields from data source like Document Type and Purchasing group. Can I select only this two field  from the list which already existing in Comm Struct and in Tran Struct.
          Does it work out and one more thing about key figures Here I need to include key figures as well......
           Please give me good advice where I come acros this issue....
    Regds
    Dayakar Reddy

  • Mapping Issue with 7 source tables and one target table in one step

    Hi,
    Request for a small help in OWB.
    I am trying to map 7 source tables to a single target table in one step.These source tables are in Oracle 10G database but dont have PK and FK relation ship.
    I am able to link one table to the target by pointing some of the columns. But when coming to the second table it is giving some error message :
    Ap18003: Connection target attribute group is already connected to a incompatible data source. Use a joiner or set operator to join the upstream data first before connection it to this operator.
    As per the error message I used a Join operator and tried to map the second table to the target but still giving me the same error message. Could some body give me a hand to give out from this step.
    Thanks for your help in advance.
    Cheers,
    Krishna.

    Hi,
    like this:
    Ingroup1
    - id -> Number(9,0)
    - name -> VARCHAR2(500)
    Ingroup2
    - my_id -> Number(9,0)
    - name -> VARCHAR2(500)
    Outgroup
    - id -> point to target_table.id
    - name -> point to target_table.name
    Not:
    Ingroup1
    - id -> Number(9,0)
    - name -> VARCHAR2(500)
    Ingroup2
    - name -> VARCHAR2(500)
    - my_id -> VARCHAR2(9)
    Regards
    Detlef
    null

  • Issue with migrating 'WI' using 'TFS migration tool' when trying migrate from '2013 to 2013'

    Hi ,
    Our final destination is to migrate projects from TFS-2010 to TFS-2013, for testing and familiarizing with the
    migration tool we trying to migrate project within the 2013 server and between two servers of 2013,
    we don't have any issues with migrating source code, we done it successfully,
    but when we try to migrate 'WI' within same server and also tried to between two server we are getting two
    different errors respectively when tried within server and between server, unable to migrate 'WI' in any scenario, log files attached for the same. Log file 'error-1' is the errors generated while trying migrate 'wI' within the same server and 'error-2' is
    the errors generated while trying migrate 'wI' between two server, Need ur guidelines to over come these issues.

    Hi Sham,
    Based on your description, I'd like to know the version of TFS integration tools you're using and the error details when migrate work items. If you use TFS integration tools 2012, please make sure you have Team Explorer 2012 installed on your server machine.
    Please also let us know the configuration you used for the migartion. If you migrate between different domains, make sure the domains have two-way trust relationship. If the issue persists, check the event logs to see if there any useful information and
    elaborate more details about your scenario with screenshots.
    Best regards,
    We are trying to better understand customer views on social support experience, so your participation in this interview project would be greatly appreciated if you have time. Thanks for helping make community forums a great place.
    Click
    HERE to participate the survey.

  • Performance issue with Oracle data source

    Hi all,
    I've a rather strange problem that I'm stuck on need some assistance on.
    I have a rules file which drags data in via an SQL data source thats an Oracle server. If I cut/paste the 3 sections of "select" "from" and "where" into SQL-Developer and run the query, it takes less than 1 second to complete. When I run the "load data" with this rule file or even use the "Retrieve" with the rules file edit, it takes up to an hour to complete/retrieve the data.
    The table in question being used has millions of rows and I'm using one of the indexed fields to retrieve the data. It's as if the Essbase/Rule file is ognoring the index, or I have a config issue with the ODBC settings on the server that is causing the problem.
    ODBC.INI file entry for the Oracle server as follows (changed any sensitive info to xxx or 999).
    [XXX]
    Driver=/opt/data01/hyperion/common/ODBC-64/Merant/5.2/lib/ARora22.so
    Description=DataDirect 5.2 Oracle Wire Protocol
    AlternateServers=
    ApplicationUsingThreads=1
    ArraySize=60000
    CachedCursorLimit=32
    CachedDescLimit=0
    CatalogIncludesSynonyms=1
    CatalogOptions=0
    ConnectionRetryCount=0
    ConnectionRetryDelay=3
    DefaultLongDataBuffLen=1024
    DescribeAtPrepare=0
    EnableDescribeParam=0
    EnableNcharSupport=0
    EnableScrollableCursors=1
    EnableStaticCursorsForLongData=0
    EnableTimestampWithTimeZone=0
    HostName=999.999.999.999
    LoadBalancing=0
    LocalTimeZoneOffset=
    LockTimeOut=-1
    LogonID=xxx
    Password=xxx
    PortNumber=1521
    ProcedureRetResults=0
    ReportCodePageConversionErrors=0
    ServiceType=0
    ServiceName=xxx
    SID=
    TimeEscapeMapping=0
    UseCurrentSchema=1
    Can anyone please advise on this lack of performance.
    Thanks in advance
    Bagpuss

    One other thing that I've seen is that if your Oracle data source and Essbase server are in different geographic locations, you can get some delay when it retrieves data over the WAN. I guess there is some handshaking going on when passing the data from Oracle to Essbase (either by record or groups of records) that is slowed WAY down over the WAN.
    Our solution to this was remove teh query out of the load rule, run it via SQL+ on a command line at the geographic location where the Oracle database is, then ftp the resulting file to where the Essbase server is.
    With upwards of 6 million records being retrieved, it took around 4 hours in the load rule, but running the query via command line took 10 minutes, then the ftp took less than 5.

  • Issue with SQL Query with Presentation Variable as Data Source in BI Publisher

    Hello All
    I have an issue with creating BIP report based on OBIEE reports which is done using direct SQL. There is this one report in OBIEE dashboard, which is written using direct SQL. To create the pixel perfect version of this report, I am creating BIP data model using SQL Query as data source. The physical query that is used to create OBIEE report has several presentation variables in its where clause.
    select TILE4,max(APPTS), 'Top Count' from
    SELECT c5 as division,nvl(DECODE (C2,0,0,(c1/c2)*100),0) AS APPTS,NTILE (4) OVER ( ORDER BY nvl(DECODE (C2,0,0,(c1/c2)*100),0))  AS TILE4,
    c4 as dept,c6 as month FROM 
    select sum(case  when T6736.TYPE = 'ATM' then T7608.COUNT end ) as c1,
         sum(case  when T6736.TYPE in ('Call Center', 'LSM') then T7608.CONFIRMED_COUNT end ) as c2,
         T802.NAME_LEVEL_6 as c3,
         T802.NAME_LEVEL_1 as c4,
         T6172.CALENDARMONTHNAMEANDYEAR as c5,
         T6172.CALENDARMONTHNUMBERINYEAR as c6,
         T802.DEPT_CODE as c7
    from
         DW_date_DIM T6736 /* z_dim_date */ ,
         DW_MONTH_DIM T6172 /* z_dim_month */ ,
         DW_GEOS_DIM T802 /* z_dim_dept_geo_hierarchy */ ,
         DW_Count_MONTH_AGG T7608 /* z_fact_Count_month_agg */
    where  ( T802.DEpt_CODE = T7608.DEPT_CODE and T802.NAME_LEVEL_1 =  '@{PV_D}{RSD}' 
    and T802.CALENDARMONTHNAMEANDYEAR = 'July 2013'
    and T6172.MONTH_KEY = T7608.MONTH_KEY and T6736.DATE_KEY = T7608.DATE_KEY
    and (T6172.CALENDARMONTHNUMBERINYEAR between substr('@{Month_Start}',0,6)  and substr('@{Month_END}',8,13))
    and (T6736.TYPE in ('Call Center', 'LSM')) )
    group by T802.DEPT_CODE, T802.NAME_LEVEL_6, T802.NAME_LEVEL_1, T6172.CALENDARMONTHNAMEANDYEAR, T6172.CALENDARMONTHNUMBERINYEAR
    order by c4, c3, c6, c7, c5
    ))where tile4=3 group by tile4
    When I try to view data after creating the data set, I get the following error:
    Failed to load XML
    XML Parsing Error: mismatched tag. Expected: . Location: http://172.20.17.142:9704/xmlpserver/servlet/xdo Line Number 2, Column 580:
    Now when I remove those Presention variables (@{PV1}, @{PV2}) in the query with some hard coded values, it is working fine.
    So I know it is the PV that's causing this error.
    How can I work around it?
    There is no way to create equivalent report without using the direct sql..
    Thanks in advance

    I have found a solution to this problem after some more investigation. PowerQuery does not support to use SQL statement as source for Teradata (possibly same for other sources as well). This is "by design" according to Microsoft. Hence the problem
    is not because different PowerQuery versions as mentioned above. When designing the query in PowerQuery in Excel make sure to use the interface/navigation to create the query/select tables and NOT a SQL statement. The SQL statement as source works fine on
    a client machine but not when scheduling it in Power BI in the cloud. I would like to see that the functionality within PowerQuery and Excel should be the same as in Power BI in the cloud. And at least when there is a difference it would be nice with documentation
    or more descriptive errors.
    //Jonas 

  • 10g Reports issue with XML Data Source

    Hi,
    Has anybody ever encountered an issue with Oracle 10g report using an XML as the data source? What happens is, some of the values in the XML are printed to the wrong column.
    One of the elements in our XML file is a complex type with 10 elements under it. The first 5 are picked up properly, but the last 6 are not. Elements #6 to #9 has a minimum occurence of 0. What happens is when element #6 is present, but #7 is, the value for element #7 is passed on to element #6.
    The XSD and XSL files are both valid since the reports were working when we were still using 9i. There is no hidden logic in the report which might cause this issue to come up, i.e., the report just picks up the values from the XML and prints it to the appropriate columns.
    Any help will be greatly appreciated.

    XSD used
    <?xml version="1.0" encoding="UTF-8"?>
    <xs:schema xmlns:xs="http://www.w3.org/2001/XMLSchema" elementFormDefault="qualified">
            <!-- trade instructions detail & trailer -->
            <xs:element name="TradeDetail">
                    <xs:complexType>
                            <xs:sequence>
                                    <xs:element ref="TradeType"/>
                                    <xs:element ref="TradeID"/>
                                    <xs:element ref="TradeDate"/>
                                    <xs:element ref="FundID"/>
                                    <xs:element ref="FundName"/>
                                    <xs:element ref="DollarValue" minOccurs="0"/>
                                    <xs:element ref="UnitValue" minOccurs="0"/>
                                    <xs:element ref="PercentageValue" minOccurs="0"/>
                                    <xs:element ref="OriginalTradeID" minOccurs="0"/>
                                    <xs:element ref="CancellationFlag"/>
                            </xs:sequence>
                    </xs:complexType>
            </xs:element>
            <xs:element name="Instruction">
                    <xs:complexType>
                            <xs:sequence minOccurs="0">
                                    <xs:element ref="TradeDetail" maxOccurs="unbounded"/>
                            </xs:sequence>
                    </xs:complexType>
            </xs:element>
            <!-- overall trade instruction message -->
            <xs:element name="InterchangeHeader">
                    <xs:complexType>
                            <xs:sequence>
                                    <xs:element ref="Instruction"/>
                            </xs:sequence>
                    </xs:complexType>
            </xs:element>
            <!-- definition of simple elements -->
            <xs:element name="FundID" type="xs:string"/>
            <xs:element name="TradeType" type="xs:string"/>
            <xs:element name="TradeID" type="xs:string"/>
            <xs:element name="TradeDate" type="xs:string"/>
            <xs:element name="FundName" type="xs:string"/>
            <xs:element name="DollarValue" type="xs:decimal"/>
            <xs:element name="UnitValue" type="xs:decimal"/>
            <xs:element name="PercentageValue" type="xs:decimal"/>
            <xs:element name="OriginalTradeID" type="xs:string"/>
            <xs:element name="CancellationFlag" type="xs:string"/>
    </xs:schema>
    XML used
    <?xml version = '1.0' encoding = 'UTF-8'?>
    <InterchangeHeader xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" xsi:noNamespaceSchemaLocation="TradeInstruction.xsd">
       <Instruction>
          <TradeDetail>
             <TradeType>Purchase</TradeType>
             <TradeID>M000038290</TradeID>
             <TradeDate>20061201</TradeDate>
             <FundID>ARO0011AU</FundID>
             <FundName>ABN Fund</FundName>
             <DollarValue>2111.53</DollarValue>
             <CancellationFlag>N</CancellationFlag>
          </TradeDetail>
          <TradeDetail>
             <TradeType>Redemption</TradeType>
             <TradeID>M000038292</TradeID>
             <TradeDate>20061201</TradeDate>
             <FundID>ARO0011AU</FundID>
             <FundName>AMRO Equity Fund</FundName>
             <UnitValue>104881.270200</UnitValue>
             <CancellationFlag>N</CancellationFlag>
          </TradeDetail>
          <TradeDetail>
             <TradeType>ISPurchase</TradeType>
             <TradeID>M000038312</TradeID>
             <TradeDate>20061201</TradeDate>
             <FundID>MLC0011AU</FundID>
             <FundName>Cash Fund</FundName>
             <OriginalTradeID>M000038311</OriginalTradeID>
             <CancellationFlag>N</CancellationFlag>
          </TradeDetail>
       </Instruction>
    </InterchangeHeader>
    XSLT used
    <?xml version="1.0" encoding="UTF-8"?>
    <xsl:stylesheet version="1.0" xmlns:xsl="http://www.w3.org/1999/XSL/Transform">
            <xsl:output method="xml" version="1.0" encoding="UTF-8" indent="yes"/>
            <xsl:template match="/">
                    <InterchangeHeader>
                            <xsl:for-each select="InterchangeHeader/Instruction/TradeDetail">
                            <xsl:sort select="FundName"/>
                            <xsl:sort select="TradeDate"/>
                                    <TradeDetail>
                                            <TradeType><xsl:value-of select="TradeType"/></TradeType>
                                            <TradeID><xsl:value-of select="TradeID"/></TradeID>
                                            <TradeDate><xsl:value-of select="TradeDate"/></TradeDate>
                                            <FundID><xsl:value-of select="FundID"/></FundID>
                                            <FundName><xsl:value-of select="FundName"/></FundName>
                                            <DollarValue><xsl:value-of select="DollarValue"/></DollarValue>
                                            <UnitValue><xsl:value-of select="UnitValue"/></UnitValue>
                                            <PercentageValue><xsl:value-of select="PercentageValue"/></PercentageValue>
                                            <OriginalTradeID><xsl:value-of select="OriginalTradeID"/></OriginalTradeID>
                                            <CancellationFlag><xsl:value-of select="CancellationFlag"/></CancellationFlag>
                                    </TradeDetail>
                            </xsl:for-each>
                    </InterchangeHeader>
            </xsl:template>
    </xsl:stylesheet>

  • Issue with Pro Res sources when encoding in Media Encoder.

    There seems to be a big issue with Pro Res sources in Media Encoder. I've noticed that when exporting using the 'software only' mode my graphics and titles look horrible, they are pixelated around the edges and the compression looks bad. This issue only happens when it's being made from a Pro Res source, if I make the exact same file Uncompressed this issue is resolved. If I use the 'Cuda' option (which I already know is the better option) this issue is resolved. The thing is, in a work environment not all of our systems are Cuda enabled and I would like to use Media Encoder as a exporting option overall. I love Media Encoder, it's fast and easy to use but this Pro Res issue is huge because the majority of the time we are working in Pro Res. I also did a test out of Avid Media Composer to Media Encoder, I sent a reference file referencing the Avid MXF material and the issue is gone, this seems to my knowledge to be a Pro Res only issue. The settings I am exporting to is 960 x 540 h.264 and also .mp4. This is coming from a 1080p source and yes I do have the 'maximum render quality' checked for best scaling. I understand that Software only vs Cuda and Open CL use different algorithms when scaling but this seems crazy to me that it would look  this much worse.
    Anyways if somebody can please look into this that would be great, this seems to be an issue where I can't continue at this moment to use Media Encoder. Making my source files Uncompressed every time to do an export is just not a real workflow I want to do.
    On a side note I've also recently noticed on a clip that had text over a grey background that there are lines going all down the screen evenly when exporting to even a non scaled 1080p .mp4. Once again, this issue goes away with any format but Pro Res. The weird part is this happens even with Cuda enabled. This is why I am thinking Media Encoder is having some sort of issue with the Pro Res codec.
    I am on a current new 27'' fully loaded iMac with the latest Adobe CC updates.
    Has anyone else experienced this?
    thank you

    No, it is why advance users do not use the wizard.
    Have a look at OpenOffice.org and its form creation. Once you add your fields in OpenOffice.org, just export to PDF and the form fields will be named just like you named in them in OpenOffice.org and the drop down values carry over.

  • Please help - Can not use stored procedure with CTE and temp table in OLEDB source

    Hi,
       I am going to create a simple package. It has OLEDB source , a Derived transformation and a OLEDB Target database.
    Now, for the OLEDB Source, I have a stored procedure with CTE and there are many temp tables inside it. When I give like EXEC <Procedure name> then I am getting the error like ''The metadata  could not be determined because statement with CTE.......uses
    temp table. 
    Please help me how to resolve this ?

    you write to the temp tables that get created at the time the procedure runs I guess
    Instead do it a staged approach, run Execute SQL to populate them, then pull the data using the source.
    You must set retainsameconnection to TRUE to be able to use the temp tables
    Arthur My Blog

Maybe you are looking for