RDW-Failed to build dimension temp tables

Hi Experts,
I am facing below issue while batch execute in RDW,
Batch: ttltypdm
Error: ttltypdm 07:29:47: ERROR: Failed to build dimension temp tables
Please Help
Regards

Hi,
can you please share the ttltypdm.ttltypdm.txt.DDMMYYYY file content or it is the one you've stated?

Similar Messages

  • SQL2005 is keeping temp table after succesfull dimension processing

    When processing dimensions, I notice that on SQL there are tmp tables that do not get deleted. These tables have a different owner (not dbo but the installation user of BPC, with the domain prefix)
    Any idea why these temp tables are there?
    Example of such a temp table: tmpMBRAccount_20100315094306   (for the dimension Account)
    Version : SAP BPC 7.0 SP7 on SQL 2005, multi server (SQL clustered - virtual app server - physical OLAP server)

    Hi Bjorn,
    Did you check the creation date of that temporary table?
    Can you reproduce the issue in Apshell?
    I was looking into my environment I don't see any table like one mentioned by you.
    To import dimension are you using a custom DTS package? This can be an explanation for that table.
    Regards
    Sorin Radulescu

  • BUILDING DIMENSION FROM ORACLE SORCE

    Hi, I am building dimensions from Oracle source using Hyperion SQL Interface. The Rule file is successfully created. However when I tried Update outline, it fails. In the log file I observe the message, Failed in initialization of log file. If anyone has faced this problem, please suggest how you could solve this? Anyone having any small documentation on using SQL interface (apart from Essbases's documentation) please forward it to my email [email protected]

    Hi,
    You can use rule files to build dimensions. You have to define the sql data source to build the dimension from the Oracle tables. Please follow given link which has all the details about building dimensions.
    http://docs.oracle.com/cd/E26232_01/doc.11122/eas_help/frameset.htm?bldrule.html
    Thanks,
    Aravindh K

  • Report with temp table

    Hi all,
    I need your suggestions.
    Have a need to build a report based on multiple queries and some updates. ( the updates are temporary, it will not update the source ). I would like to put data into a temporary table that is good for the log on session only then build the report based on this temporary table. Can BIP handle this ? I am open to any ideas. by the way report will be called from APEX side.
    oralce 10G, apex v4
    thx in advance.
    tai

    I set the data parameter at the report level as a DateTime type. It is defaulted to today()
    Only dataset1 uses the date parameter, dataset 2 doesn't.
    The query uses
    WHERE tdate =  ISNULL(@T_Date, CONVERT(VARCHAR(8), GETDATE(), 1))
    If I get rid of Dataset2, then the report is OK, so Dataset1 is able to read the date. But if Dataset2 which uses the temp table is added, then I get this error
             An error occured during local report processing
             An error occured during report processing
    Query execution failed for dataset Dataset2
    Invalid object name #temptable

  • Global Temp Table or Permanent Temp Tables

    I have been doing research for a few weeks and trying to comfirm theories with bench tests concerning which is more performant... GTTs or permanent temp tables. I was curious as to what others felt on this topic.
    I used FOR loops to test out the performance on inserting and at times with high number of rows the permanent temp table seemed to be much faster than the GTTs; contrary to many white papers and case studies that have read that GTTs are much faster.
    All I did was FOR loops which iterated INSERT/VALUES up to 10 million records. And for 10 mil records, the permanent temp table was over 500k milliseconds faster...
    Anyone have an useful tips or info that can help me determine which will be best in certain cases? The tables will be used for staging for ETL Batch processing into a Data Warehouse. Rows within my fact and detail tables can reach to the millions before being moved to archives. Thanks so much in advance.
    -Tim

    > Do you have any specific experiences you would like to share?
    I use both - GTTs and plain normal tables. The problem dictates the tools. :-)
    I do have an exception though that does not use GTTs and still support "restartability".
    I need to to continuously roll up (aggregate) data. Raw data collected for an hour gets aggregated into an hourly partition. Hourly partitions gets rolled up into a daily partition. Several billion rows are processed like this monthly.
    The eventual method I've implemented is a cross between materialised views and GTTs. Instead of dropping or truncating the source partition and running an insert to repopulate it with the latest aggregated data, I wrote an API that allows you to give it the name of the destination table, the name of the partition to "refresh", and a SQL (that does the aggregation - kind of like the select part of a MV).
    It creates a brand new staging table using a CTAS, inspects the partitioned table, slaps the same indexes on the staging table, and then performs a partition exchange to replace the stale contents of the partition with that of the freshly build staging table.
    No expensive delete. No truncate that results in an empty and query-useless partition for several minutes while the data is refreshed.
    And any number of these partition refreshes can run in parallel.
    Why not use a GTT? Because they cannot be used in a partition exchange. And the cost of writing data into a GTT has to be weighed against the cost of using that data by writing it (or some of it) into permanent tables. Ideally one wants to plough through a data set once.
    Oracle has a fairly rich feature set - and these can be employed in all kinds of ways to get the job done.

  • Global Temp Table Not found - SSIS

    I am facing below error while using global temp table in SSIS.
    [OLE DB Destination [78]] Error: SSIS Error Code DTS_E_OLEDBERROR.  An OLE DB error has occurred. Error code: 0x80040E37.
    An OLE DB record is available.  Source: "Microsoft SQL Server Native Client 10.0"  Hresult: 0x80040E37  Description: "Table/view either does not exist or contains errors.".
    [OLE DB Destination [78]] Error: Failed to open a fastload rowset for " ##AGENTDTLS". Check that the object exists in the database.
    [SSIS.Pipeline] Error: component "OLE DB Destination" (78) failed the pre-execute phase and returned error code 0xC0202040.
    1) For data connection manager - Retain same connection is set to True
    2) Data Flow task - Delay Validation is set to True
    3) Destination Task - Using Temp Table - ValidateExternalMetadata is set to false.
    4) I am just using one data connection.
    5) before using the temp file I am checking if its exits and if yes drp it first and create it.
    Not able to understand the reason for failure.

    Why don't you use permanent table in tempdb?
    Kalman Toth Database & OLAP Architect
    SQL Server 2014 Design & Programming
    New Book / Kindle: Exam 70-461 Bootcamp: Querying Microsoft SQL Server 2012

  • Use of Oracle global temp table in BI Publisher

    Hi All,
    I have witten a function which populates a Gobal temp table.I call the function and do a select on the global temp table.This fails to give me any result.
    We are calling the function within the package in the 'beforeReport' trigger in our Data Template in BI Publisher.
    I tried to check the session id and found that the sessions are different when we execute the functiona and when we do a select query.
    How can I avoid this situation of changing session?
    I have tried by removing commit from the function but even thta didnt help.
    Please let me know if anyone has suugestion on this.
    Thanks in advance!

    Hi,
    To check if the function is working fine , we inserted the output from temp table into a normal physical table .We queried the physical table and found the values there.
    So , the function is working fine.
    We dont have any statement like 'PRAGMA AUTONOMOUS_TRANSACTION ' in our function.
    Do we need to add it?
    Just to clarify the issue , We have populated the temp table by calling a function and after that we are trying to query the temp table.However, it appears that the session changes between teh function call and issuing of select query .This is resultin into no data being fetched as the values in Gloabal temp table persist only for a session.
    Thanks!

  • How do I CREATE IF NOT EXISTS Temp table in PLSQL?

    hello, how do I CREATE IF NOT EXISTS Temp table in PLSQL? The following table is to be created in FIRST call inside a recursive function (which you'll see in QUESTION 2).
    QUESTION 1:
    CREATE GLOBAL TEMPORARY TABLE TmpHierarchyMap
                                  Id numeric(19,0) NOT NULL,
                                  ParentId numeric(19,0) NOT NULL,
                                  ChildId numeric(19,0) NOT NULL,
    ... more ...
                             ) on commit delete rows');
    QUESTION 2: How to return a temp table from a function?
    For example, this is how I'm doing it at the moment, using Nested Table.
    EXECUTE IMMEDIATE 'CREATE OR REPLACE TYPE TmpHierarchyMapObjType AS OBJECT
                   Id numeric(19,0) ,
                   ParentId numeric(19,0),
                   ChildId numeric(19,0),
    ... more ...
         EXECUTE IMMEDIATE 'CREATE OR REPLACE TYPE TmpHierarchyMapTableType AS TABLE OF TmpHierarchyMapObjType;';
    CREATE OR REPLACE FUNCTION fnGetParentsTable
    ObjectId number,
    ObjectClassifier varchar2
    RETURN TmpHierarchyMapTableType
    IS
    TmpHierarchyMap TmpHierarchyMapTableType := TmpHierarchyMapTableType();
    ThisTempId varchar2(32);
    CURSOR spGetParents_cursor IS
    SELECT
    Id,
    ParentId,
    ChildId,
    FROM TMP_HIERARCHYMAP
    WHERE TempId = ThisTempId;
    BEGIN
    SELECT sys_guid() INTO ThisTempId FROM dual;
    spRecursiveGetParents(ObjectId, ObjectClassifier, ThisTempId);
    FOR oMap in spGetParents_cursor LOOP
    TmpHierarchyMap.Extend();
    TmpHierarchyMap(TmpHierarchyMap.Count) := TmpHierarchyMapObjType( oMap.Id
    , oMap.ParentId
    , oMap.ChildId
    END LOOP;
    DELETE FROM TMP_HIERARCHYMAP WHERE TempId = ThisTempId;
    RETURN TmpHierarchyMap;
    END fnGetParentsTable;
    QUESTION 3: what does the word GLOBAL means? I read that temp table is visible only to a particular database connection/session and will be dropped automatically on termination of the session. i can only find this information in some forum discussion but failed to locate this in Oracle doc, can someone point me in right direction please?
    Many thanks!
    REF:
    http://stackoverflow.com/questions/221822/sybase-developer-asks-how-to-create-a-temporary-table-in-oracle
    http://www.oracle-base.com/articles/8i/TemporaryTables.php

    devvvy wrote:
    so if I CREATE GLOBAL TEMPORARY TABLE twice on second pass of my recursive function what then...?You don't create it inside your function.
    You create the GTT once on your database outside of the function and then just leave it there and use it.
    Tables should not be dynamically created and dropped.
    Only other database engines such as SQL Server use the concept of creating temporary tables during the execution of code. Oracle uses a "create once, use whenever" methodology.

  • Issue with Temp tables in SSIS 2012 with RetainSameConnection=true

    Hello,
    We have few packages written in 2008 and are being upgraded to 2012. Our package mostly uses temp tables during the process.  During initial migration, we faced issue with handling temp table in the OLE Db destination provider and found a solution for
    the same under 
    usage of Temp tables in SSIS 2012
    Most of our packages execute fine now. 
    we came across a different issue recently. For one of our package, which merges 3 feeds into a temp table and then executes a stored procedure for processing, the package fails intermittently.
    Below are properties of SSIS and its components, which you might be interested
    * Retainsameconnection for the OLE Db connection manager set to True
    * properties of OLEDB Destination 
    AccessMode : SQL Command
    CommandTimeOut : 0
    SQLCommand : Select * from #tmp
    * using SSIS 2012 and SQL OLEDB Native Provider 11 (Provider=SQLNCLI11.1)
    * one of the feed is 10MB
    During investigation using profiler, found that though I use RetainSameConnection, I often  could see that more than one SPId is used during the scope of SSIS execution and when ever this happens, package fails with below error message 
    An OLE DB record is available. Source: "Microsoft SQL Server Native Client 11.0" Hresult: 0x80040E14 Description: "Statement(s) could not be prepared.".
    An OLE DB record is available. Source: "Microsoft SQL Server Native Client 11.0" Hresult: 0x80040E14 Description: "Invalid object name '#tmp'."
    Now, Why SSIS uses a different SPId during its execution when RetainSameconnection is set to True (note : I have only one OLEDB connection in that package)? 
    To simulate the issue, Instead of 10MB file, I used a 500KB file and execute the package twice and all went fine.
    Is it because of 10 MB file taking long time to process causing the time out of that OLEDB destionation forcing the SSIS to go for another connection? but remember, CommandTimeout is set to infinite(0) for that OLEDB destination. 
    Much appreciated your response. 

    Hey,
    I understand you used Retainsameconnection property true for all the OLEDB connections you used in the package if not make sure its for all the connection including file connection as well.
    Additionally, you can try to set Delayvalidation property to true for all the dataflows and control flows in the connection and try running the package for 10MB file.
    I hope this will fix the intermittent failure issue you are facing with SSIS.
    (Please mark solved if I've answered your question, vote for it as helpful to help other user's find a solution quicker)
    Thanks,
    Atul Gaikwad.

  • How to build semantic model table from excel 2013 budget table?

    I have an excel 2013 budget table produced by finance department. The title of the table is <<Year 2015 Resource Budget>> with columns such as (Item, Budget). There is line of business app which tracks daily usage. It is SQL database which
    I can access. The SQL table has columns such as Date, Item, Usage. I am asked to produce some PowerView reports using DAX.
    I am thinking about using Excel 2013 PowerPivot. The following is the step of my plan.
    Add excel 2013 Budget table to PowerPivot
    Transform the budget table using DAX to a new table with includes date. For each item budget, the value can be evenly divided by 365. I don't know if this is possible with DAX and need advice from you
    Import daily usage from SQL database into PowerPivot
    Merge the imported SQL daily usage table with the transformed budget table with daily breakdown.  The output table has 4 columns (Date, Item, ValueType, Value). The ValueType can be either Budget and Usage. (I also do not know if it is possible to
    use DAX to merge two PowerPivot table into anther table with added new columns. If it is possible, I would love to hear from you on how to do this)
    Finally, I can produce BI reports based on the newly merged table
    Update: Here are more sample data.
    Input Budget
    <<Year 2014 Resource Budget>>
    Item
    Budge
    Pen
    365
    Paper
    1095
    <<Year 2015 Resource Budget>>
    Item
    Budge
    Pen
    1825
    Paper
    1095
    Bag
    365
    I am thinking to expand the Budget as follow but I don't how to use DAX to do it
    Date
    Item
    Budge
    31/12/2014
    Pen
    1
    31/12/2014
    Paper
    3
    1/01/2015
    Pen
    5
    1/01/2015
    Paper
    3
    1/01/2015
    Bag
    1
    Here is the Daily Usage table
    Date
    Item
    Usage
    31/12/2014
    Paper
    2
    31/12/2014
    Pen
    6
    1/01/2015
    Paper
    10
    1/01/2015
    Pen
    2
    1/01/2015
    Bag
    1
    I was thinking to merge the expanded Daily Budget and Daily Usage as follow for BI reporting. This was my thinking, may not a good solution using DAX
    Date
    Item
    Type
    Value
    31/12/2014
    Paper
    Usage
    2
    31/12/2014
    Pen
    Usage
    6
    1/01/2015
    Paper
    Usage
    10
    1/01/2015
    Pen
    Usage
    2
    1/01/2015
    Bag
    Usage
    1
    31/12/2014
    Pen
    Budget
    1
    31/12/2014
    Paper
    Budget
    3
    1/01/2015
    Pen
    Budget
    5
    1/01/2015
    Paper
    Budget
    3
    1/01/2015
    Bag
    Budget
    1
    Off course, I need to build dimension tables for Date, Item, Item Category, and Value Type
    jl

    Hi JL,
    Just for consideration, I'd:
    Keep budget and actuals separately - create Fact_Y2015Budget and Fact_Actuals;
    Hook both tables to same dimesnions (time; if there is no time for budget, assign it all to the first day of your 2015; line item etc. - whatever you are asked to report about)
    If you can, pull usage not only for 2015, but for prior year(s) too
    Create measures like [Usage Budget Atteinment YTD]:=TotalYTD(SUM([Usage]),your_time)/([Budget]*PriorYear(TotaYTD(SUM([Usage]),you_time)/CALCULATE(SUM([Usage]),DATESBETWEEN(your_first_day_of_year2015,your_last_day_of_2015))))
    The calculation in #4 compares actual ("usage") to portion of budget that would have been used up if 2015 had the same seasonality as 2014... If you couldmshare specific sample, I'd buuild exact formula...
    Daniel

  • Urgent!  Slow Result Set -- temp table slowing me??

    -- Running BC4J/JHeadstart/UIX
    Description:
    I have a uix page that calls a Servlet and passes a TABLE_NAME. The Servlet gets the TABLE_NAME and calls a class that extends oracle.jheadstart.persistence.bc4j.handler.DataSourceHandlerImpl to create a ViewObject and get the data we need. Once the ViewObject is passed back to the servlet, the servlet loops through the VIewObject and builds a report. See the problem below and the code at the bottom.
    Problem:
    I am running a query that returns approx 5000 records to my ViewObject. I then loop through the rows and construct a report. The view object will return the first 1085 records quickly, however, the following 4000 come back very slowly. I read online that BC4J creates temp tables to store large resultsets and then streams the data to the user as you need it. Is this our potential bottleneck?
    Questions:
    Is there a way to have it return all the rows? What can I do to speed this up?
    Code:
    --- Begin Servlet Snippet ---
    private ByteArrayOutputStream createReport(HttpServletRequest request)
    try{
    // PARM_REPORT = table name
    String reportName = request.getParameter(PARM_REPORT);
    System.out.println(">> calling getReport for " + reportName);
    RdmUserHandlerImpl handler = new RdmUserHandlerImpl();
    ViewObject vo = handler.getReportView2(reportName, request.getSession().getId());
    System.out.println(">> back from get report");
    // loop through report and print every 100
    while(vo.hasNext())
    curRow++;
    if (curRow % 100 == 0 )
    System.out.println(curRow + "");
    --- End Servlet Snippet ---
    --- Begin RdmUserHandlerImpl Snippet ---
    public ViewObject getReportView2(String tableName, Object sessionId) throws Exception {
    System.out.println("IN GET REPORT VIEW");
    ApplicationModule appMod = (ApplicationModule)getConnection("classpath...resource.MyUser", sessionId);
    // First see if we already created the view definition
    ViewObject vo = appMod.findViewObject(tableName);
    // If it was already created then refresh it, else lets try to create it
    if(vo != null) {
    System.out.println("found existing view");
    vo.reset();
    else {
    System.out.println("view not found, making new view");
    String query="SELECT * FROM " + tableName;
    System.out.println("QUERY = " + query);
    vo = appMod.createViewObjectFromQueryStmt(tableName, query);
    // max fetch returns -1
    System.out.println("MAX Fetch Size = " + vo.getMaxFetchSize());
    return vo;
    --- End RdmUserHandlerImpl Snippet ---
    Please reply asap! Deadline is coming fast!
    -Matt

    Matt,
    I think that you are right, the temporary tables created by BC4J are the reason for making it slow after a certain number of records. One of Steve Muench's articles includes the text:
    One of the most frequent performance-related questions we get on the Oracle Technet discussion forum is a question like, "After I query about a 1000 rows in a view object, my application gets very, very slow. What's happening?"
    It explains how you can turn off this feature, see the full article at http://www.oracle.com/technology/products/jdev/tips/muench/voperftips/index.html.
    The following article gives a lot of helpful information about the temporary tables:
    http://www.oracle.com/technology/products/jdev/htdocs/bc4j/bc4j_temp_tables.html
    The next article gives general tips for performance tuning of BC4J:
    http://www.oracle.com/technology/products/jdev/howtos/10g/adfbc_perf_and_tuning.html
    Hope this helps,
    Sandra Muller
    JHeadstart Team

  • Are there any (dis)advantages in building a universe on fully normalized tables instead of building dimensional model/tables and then universe on top of them?

    Hello,
    I’m hoping someone can help me with understanding advantages and disadvantages if we want to build a universe on top of a fully normalized tables, compared to using a dimensional model with star schemas.
    I’ve read some discussions here that say that it is possible to create a universe on top of normalized tables. Then, can we avoid building of dimensional tables (a data mart), and just use normalized tables? I would say that it is easier to use star schema dimensions and facts tables to build a universe, but our end users might ask “why do we have to go through building a dimensional data mart, if we can have same reports with hierarchies and drill-down functionality based on a universe built on top of our already existing normalized tables?”
    Can you point me to some established best practices regarding using normalized tables to build a universe? Any documents with some examples for this?
    Any expected difficulties during design/development phase of our universe, related to using normalized tables?
    Any expected performance degradation if we use normalized tables compared to using dimensional tables?
    Can I build universe more easily if I transform (modify) our normalized model (by using alias tables and views) to look like snowflake model?
    I’m using BOE XI 3.1, tables are in Oracle 11.2.
    Thank you

    Few Disadvantages that I usually face when building universe on Normalized tables,
    1. Performance - Read operations have to suffer because indexing strategies do not go well with table joins
    2. Derived Tables - Due to complex Queries/ Logic, most of the time, end-up creating derived tables, which doesn't use back-end table indexes and slows down the report runtime.
    3. Normalized table/ Transaction tables may not always have proper cardinality and often results in Cartesian products
    4. Normalized tables may not have tight referential integrity and may have to join more than one column and join on varchar, etc whereas, good Dimensional model datawarehouse will have proper keys/ integer joins and not always necessary to join on multiple fields
    5. Often deal with Fan and Chasm Traps
    6. Dealing with Database fields with nulls, blanks, date in numeric format, etc.,
    7. No Facts, Dimensions separated and most of the time they are in same table
    and More...
    If performance is not a matter and building Datawarehouse is a big deal, then I will start building Universe on normalized tables by having the database diagram as reference for Joins, contexts, etc
    Note: After dealing with universes based on normalized tables for few years (by creating views, complex sql queries data loading to tables and unv on these tables, derived tables), I ended up creating star-schema dimensional model (couple of months extra ETL work), users/ developers felt lot better when they have to create standard/ ad-hoc reports and they are super fast compared to previous universes.

  • What to fill in "temp table scope" for global temporary tables?

    Hi,
    I'm using Data Modeler 4.0.1.836 and whatever I put in the "temp table scope" box for a global temporary table doesn't seem to affect the DDL script regarding the ON COMMIT PRESERVE/DELETE ROWS option. The script always shows ON COMMIT PRESERVE ROWS no matter what.
    Yet, some of my temporary tables must be created as ON COMMIT DELETE ROWS.
    The Data Modeler help says the following about this :
    Temp Table Scope:
    For a table classified as Temporary, you can specify a scope, such as Session or Dimension.
    Not sure what "Dimension" has to do with the scope here, but it doesn't make any difference.
    I tried putting "Session", "Dimension", "Transaction", but no luck. So what's the text to put for the script to generate ON COMMIT DELETE ROWS?
    Thanks

    Hi,
    The Temporary Table Scope property (on the Classification Types page of the Table Properties dialog) is purely documentary.
    To set ON COMMIT DELETE ROWS you should expand the Browser node for the Relational Model and look for the node for the relevant Oracle Physical Model.  If you expand this you will find an entry there for your Table. Double-click on this to get the Physical Model properties dialog for your table, and you will find a "Temporary"  property which has options YES (Preserve Rows), YES (Delete Rows) or NO.
    David

  • MS SQL Server 2014: Error inserting into Temp table with index and identity field

    In this thread, I mentioned a problem with SQL Server 2014:
    SQL Server 2014: Bug with IDENTITY INSERT ON
    The question was answered, it is a bug. To keep you informed on this issue, I open this discussion.
    Problem:
    The code below works perfectly fine on MS SQL Server 2008 R2 and MS SQL Server 2012, but gives an error every second time the proc is executed on MS SQL Server 2014. If I do not define any index on the temp table, the problem disappears. Defining the index
    after the insert, does not help.
    SET NOCOUNT ON
    GO
    IF EXISTS (SELECT 1 FROM sys.procedures WHERE name = 'usp_Test') DROP PROC dbo.usp_Test;
    GO
    CREATE PROC dbo.usp_Test AS
    BEGIN
    SET NOCOUNT ON
    CREATE TABLE #Source(ID integer NOT NULL);
    INSERT INTO #Source VALUES (1), (2), (3);
    CREATE TABLE #Dest (ID integer IDENTITY(1,1) NOT NULL);
    CREATE INDEX #IDX_Dest ON #Dest (ID);
    PRINT 'Check if the insert might cause an identity crisis';
    SELECT 'Source' AS SourceTable, * FROM #Source;
    SELECT 'Destination' AS DestTable, * FROM #Dest;
    SET IDENTITY_INSERT #Dest ON;
    PRINT 'Do the insert';
    INSERT INTO #Dest (ID) SELECT ID FROM #Source;
    PRINT 'Insert ready';
    SET IDENTITY_INSERT #Dest OFF;
    SELECT * FROM #Dest;
    DROP TABLE #Source;
    DROP TABLE #Dest;
    END;
    GO
    PRINT 'First execution of the proc, everything OK';
    EXEC dbo.usp_Test;
    PRINT '';
    PRINT 'Second execution of the proc, the insert fails.';
    PRINT 'Removing the index #IDX_Dest causes the error to disappear.';
    EXEC dbo.usp_Test;
    GO
    DROP PROC dbo.usp_Test;
    GO

    There is some progress. Communication from a former Microsoft employee tells us this:
    Shivendra Vishal
    Engineer at Microsoft
    I am no longer with MS, and I do not have code access, however from the public symbols, I could make out following:
    sqlmin!SetidentI2I4+0x1f3:
    000007fe`f4d865d3 488b10 mov rdx,qword ptr [rax] ds:00000000`00000000=????????????????
    ExceptionAddress: 000007fef4d865d3 (sqlmin!SetidentI2I4+0x00000000000001f3)
    ExceptionCode: c0000005 (Access violation)
    ExceptionFlags: 00000000
    NumberParameters: 2
    Parameter[0]: 0000000000000000
    Parameter[1]: 0000000000000000
    Attempt to read from address 0000000000000000
    This is a read AV and from registers it is clear that we were trying to move the value of location pointed by qword of register rax which is not valid:
    rax=0000000000000000 rbx=0000000000000038 rcx=0000000000001030
    rdx=0000000000000006 rsi=00000001f55def98 rdi=00000000106fd070
    rip=000007fef4d865d3 rsp=00000000106fcf40 rbp=00000000106fcfe9
    r8=0000000000000000 r9=00000001f55def60 r10=00000001f55defa0
    r11=00000000106fcd20 r12=0000000000000000 r13=0000000000000002
    r14=00000001f49c3860 r15=00000001f58c0040
    iopl=0 nv up ei pl nz na po nc
    cs=0033 ss=002b ds=002b es=002b fs=0053 gs=002b efl=00010206
    The stack is:
    # Child-SP RetAddr Call Site
    00 00000000`106fcf40 000007fe`f30c1437 sqlmin!SetidentI2I4+0x1f3
    01 00000000`106fd050 000007fe`f474e7ce sqlTsEs!CEsExec::GeneralEval4+0xe7
    02 00000000`106fd120 000007fe`f470e6ef sqlmin!CQScanUpdateNew::GetRow+0x43d
    03 00000000`106fd1d0 000007fe`f08ff517 sqlmin!CQueryScan::GetRow+0x81
    04 00000000`106fd200 000007fe`f091cebe sqllang!CXStmtQuery::ErsqExecuteQuery+0x36d
    05 00000000`106fd390 000007fe`f091ccb9 sqllang!CXStmtDML::XretDMLExecute+0x2ee
    06 00000000`106fd480 000007fe`f08fa058 sqllang!CXStmtDML::XretExecute+0xad
    07 00000000`106fd4b0 000007fe`f08fb66b sqllang!CMsqlExecContext::ExecuteStmts<1,1>+0x427
    08 00000000`106fd5f0 000007fe`f08fac2e sqllang!CMsqlExecContext::FExecute+0xa33
    09 00000000`106fd7e0 000007fe`f152cfaa sqllang!CSQLSource::Execute+0x86c
    0a 00000000`106fd9b0 000007fe`f152c9e8 sqllang!CStmtExecProc::XretLocalExec+0x25a
    0b 00000000`106fda30 000007fe`f152a1d8 sqllang!CStmtExecProc::XretExecExecute+0x4e8
    0c 00000000`106fe1e0 000007fe`f08fa058 sqllang!CXStmtExecProc::XretExecute+0x38
    0d 00000000`106fe220 000007fe`f08fb66b sqllang!CMsqlExecContext::ExecuteStmts<1,1>+0x427
    0e 00000000`106fe360 000007fe`f08fac2e sqllang!CMsqlExecContext::FExecute+0xa33
    0f 00000000`106fe550 000007fe`f0902267 sqllang!CSQLSource::Execute+0x86c
    10 00000000`106fe720 000007fe`f0909087 sqllang!process_request+0xa57
    11 00000000`106feee0 000007fe`f2bf49d0 sqllang!process_commands+0x4a3
    12 00000000`106ff200 000007fe`f2bf47b4 sqldk!SOS_Task::Param::Execute+0x21e
    13 00000000`106ff800 000007fe`f2bf45b6 sqldk!SOS_Scheduler::RunTask+0xa8
    14 00000000`106ff870 000007fe`f2c136ff sqldk!SOS_Scheduler::ProcessTasks+0x279
    15 00000000`106ff8f0 000007fe`f2c138f0 sqldk!SchedulerManager::WorkerEntryPoint+0x24c
    16 00000000`106ff990 000007fe`f2c13246 sqldk!SystemThread::RunWorker+0x8f
    17 00000000`106ff9c0 000007fe`f2c13558 sqldk!SystemThreadDispatcher::ProcessWorker+0x3ab
    18 00000000`106ffa70 00000000`775d59ed sqldk!SchedulerManager::ThreadEntryPoint+0x226
    19 00000000`106ffb10 00000000`7780c541 kernel32!BaseThreadInitThunk+0xd
    1a 00000000`106ffb40 00000000`00000000 ntdll!RtlUserThreadStart+0x21
    Unassembling the function:
    000007fe`f4d8658e 4c8b10 mov r10,qword ptr [rax]
    000007fe`f4d86591 4533e4 xor r12d,r12d
    000007fe`f4d86594 410fb7d5 movzx edx,r13w
    000007fe`f4d86598 4533c9 xor r9d,r9d
    000007fe`f4d8659b 4533c0 xor r8d,r8d
    000007fe`f4d8659e 488bc8 mov rcx,rax
    000007fe`f4d865a1 4489642420 mov dword ptr [rsp+20h],r12d
    000007fe`f4d865a6 41ff5230 call qword ptr [r10+30h]
    000007fe`f4d865aa 8b5597 mov edx,dword ptr [rbp-69h]
    000007fe`f4d865ad 4c8b10 mov r10,qword ptr [rax]
    000007fe`f4d865b0 4489642438 mov dword ptr [rsp+38h],r12d
    000007fe`f4d865b5 4489642430 mov dword ptr [rsp+30h],r12d
    000007fe`f4d865ba 458d442401 lea r8d,[r12+1]
    000007fe`f4d865bf 4533c9 xor r9d,r9d
    000007fe`f4d865c2 488bc8 mov rcx,rax
    000007fe`f4d865c5 c644242801 mov byte ptr [rsp+28h],1
    000007fe`f4d865ca 4488642420 mov byte ptr [rsp+20h],r12b
    000007fe`f4d865cf 41ff5250 call qword ptr [r10+50h]
    000007fe`f4d865d3 488b10 mov rdx,qword ptr [rax] <=================== AV happened over here
    000007fe`f4d865d6 488bc8 mov rcx,rax
    000007fe`f4d865d9 4c8bf0 mov r14,rax
    000007fe`f4d865dc ff5268 call qword ptr [rdx+68h]
    000007fe`f4d865df 488d55e7 lea rdx,[rbp-19h]
    000007fe`f4d865e3 4c8b00 mov r8,qword ptr [rax]
    000007fe`f4d865e6 488bc8 mov rcx,rax
    000007fe`f4d865e9 41ff5010 call qword ptr [r8+10h]
    000007fe`f4d865ed f6450a04 test byte ptr [rbp+0Ah],4
    I remember few issues with scan2ident function, I am not sure if they have fixed it however it appears that this is intoduced to SQL 2014 and we need help from MS to get this resolved as it needs code analysis.
    It is not getting simulated for other versions of SQL apart from SQL 2014.
    Also to add, interestingly, the value of rax is not visibly changed and it was successfully passed on to rcx, which has a valid value, so something should have changed the value of rax inside call to function using call qword ptr [r10+50h], and looking at this
    it appears that it might be a list of functions and we are going at particular offset [50h]. So, bottom line is that the call to function qword ptr [r10+50h], should be changing something in rax, and debugging/analyzing this code might give us some more idea.

  • XSS SCA import in CMS error: failed to build

    We are getting error when trying to build XSS Track in NWDI. There are 2 brocken DC for development track.
    Here is the error message that we get:
    Development Component Build (2008-11-28 22:04:10)
      Component name: mss/mol
      Component vendor: sap.com
      SC compartment: sap.com_SAP_MSS_1
      Configuration: <SID>_XSS_D
      Location: <SID>_XSS_D
      Source code location: http://server01:55100/dtr/ws/XSS/sap.com_SAP_MSS/dev/active/DCs/sap.com/mss/mol/_comp/
      DC root folder: /usr/sap/<SID>/JC51/j2ee/cluster/server0/temp/CBS/27/.B/6448/DCs/sap.com/mss/mol/_comp/
      DC type: Web Dynpro
      Host: server01
    DC Model check:
       All used DCs are available locally
       validating dependency to build plugin "sap.com/tc/bi/bp/webDynpro"
       validating dependency to  public part "FloorplanManager" of DC "sap.com/pcui_gp/xssfpm"
       validating dependency to  public part "default" of DC "sap.com/tc/wd/webdynpro"
       validating dependency to  public part "default" of DC "sap.com/com.sap.exception"
       validating dependency to  public part "default" of DC "sap.com/tc/ddic/ddicruntime"
       validating dependency to  public part "default" of DC "sap.com/tc/col/api"
       validating dependency to  public part "default" of DC "sap.com/tc/wdp/metamodel/content"
       validating dependency to  public part "Isrlpd" of DC "sap.com/pcui_gp/isrlpd"
       validating dependency to  public part "default" of DC "sap.com/com.sap.mw.jco"
       validating dependency to  public part "default" of DC "sap.com/com.sap.aii.proxy.framework"
       validating dependency to  public part "default" of DC "sap.com/tc/logging"
       validating dependency to  public part "default" of DC "sap.com/com.sap.aii.util.misc"
       validating dependency to  public part "FPMUtils" of DC "sap.com/pcui_gp/xssutils"
       validating dependency to  public part "default" of DC "sap.com/tc/ddic/metamodel/content"
       validating dependency to  public part "default" of DC "sap.com/tc/cmi"
       validating dependency to  public part "FinUtils" of DC "sap.com/mss/finutils"
       DC model check OK
    Start build plugin:
       using build plugin: sap.com/tc/bi/bp/webDynpro
       starting build plugin from : /usr/sap/<SID>/JC51/j2ee/cluster/server0/temp/CBS/27/.B/6448/DCs/sap.com/tc/bi/bp/webDynpro/_comp/gen/default/public/webDynpro/
    Build Plugin 'WebdynproPlugin', Version 7.00 SP 16 (645_VAL_REL, built on 2008-06-09 20:46:40 CEST, CL83275)
       development component:  mss/mol (sap.com)
          software component:  SAP_MSS (sap.com)
           support component:  EP-PCT-MGR-CO
                    location:  <SID>_XSS_D
                        type:  Web Dynpro
               build variant:  default
             output location:  /usr/sap/<SID>/JC51/j2ee/cluster/server0/temp/CBS/27/.B/6448/DCs/sap.com/mss/mol/_comp/gen/default
           generation folder:  /usr/sap/<SID>/JC51/j2ee/cluster/server0/temp/CBS/27/.B/6448/t/64CE6536D7BD962CDCB65DC8A7640047
           plugin start time:  2008-11-28 22:04:11 GMT-07:00 (MST)
                     Java VM:  Java HotSpot(TM) Client VM, 1.4.2_12-b03 (Sun Microsystems Inc.)
    General options:
      convert *.xlf to *.properties: yes
      include sources for debugging: yes
    Warning: Source folder "META-INF" exists but is empty and will be ignored.
    Warning: Source folder "src/configuration" exists but is empty and will be ignored.
    Preparing data context..
    No public part descriptor found for component "tc/wd/webdynpro" (vendor "sap.com"), public part "default", using legacy mode.
    No public part descriptor found for component "com.sap.exception" (vendor "sap.com"), public part "default", using legacy mode.
    No public part descriptor found for component "tc/ddic/ddicruntime" (vendor "sap.com"), public part "default", using legacy mode.
    No public part descriptor found for component "tc/col/api" (vendor "sap.com"), public part "default", using legacy mode.
    No public part descriptor found for component "tc/wdp/metamodel/content" (vendor "sap.com"), public part "default", using legacy mode.
    No public part descriptor found for component "com.sap.mw.jco" (vendor "sap.com"), public part "default", using legacy mode.
    No public part descriptor found for component "com.sap.aii.proxy.framework" (vendor "sap.com"), public part "default", using legacy mode.
    No public part descriptor found for component "tc/logging" (vendor "sap.com"), public part "default", using legacy mode.
    No public part descriptor found for component "com.sap.aii.util.misc" (vendor "sap.com"), public part "default", using legacy mode.
    No public part descriptor found for component "tc/ddic/metamodel/content" (vendor "sap.com"), public part "default", using legacy mode.
    No public part descriptor found for component "tc/cmi" (vendor "sap.com"), public part "default", using legacy mode.
    No 'default' JDK defined, will use running VM.
    Data context preparation finished in 0.934 seconds
    Creating Ant build file..
      Creating output file: /usr/sap/<SID>/JC51/j2ee/cluster/server0/temp/CBS/27/.B/6448/DCs/sap.com/mss/mol/_comp/gen/default/logs/build.xml
      Using macro file:     /usr/sap/<SID>/JC51/j2ee/cluster/server0/temp/CBS/27/.B/6448/DCs/sap.com/tc/bi/bp/webDynpro/_comp/gen/default/public/webDynpro/macros/build.vm
    Build file creation finished in 1 second
    Starting Ant..
      Using build file:     /usr/sap/<SID>/JC51/j2ee/cluster/server0/temp/CBS/27/.B/6448/DCs/sap.com/mss/mol/_comp/gen/default/logs/build.xml
      Using build target:   build
      Using Ant version:    1.6.2
    compile:
          [echo] Starting dictionary generator
         [ddgen]
         [ddgen] [Info]    Property deployment is true: Deployment information is provided!
         [ddgen] [Info]    Property sourcepath: /usr/sap/<SID>/JC51/j2ee/cluster/server0/temp/CBS/27/.B/6448/DCs/sap.com/mss/mol/_comp/src/packages
         [ddgen] [Info]    Property targetpath: /usr/sap/<SID>/JC51/j2ee/cluster/server0/temp/CBS/27/.B/6448/t/64CE6536D7BD962CDCB65DC8A7640047/gen_ddic
         [ddgen] [Info]    Property archivename: sap.commssmol
         [ddgen] [Info]    Property vendor: sap.com
         [ddgen] [Info]    Property dcname: mss/mol
         [ddgen] [Info]    Property projectlanguage: de
         [ddgen] [Info]    Property language: Available languages are automatically determined!
         [ddgen] [Info]    Property addpaths ...
         [ddgen] [Info]    Generating datatypes/com/sap/mss/mol/monitor/model/types/MolModel.gdictionary
         [ddgen] [Info]    Generating datatypes/com/sap/pcuigp/xssutils/objectselection/model/types/FcObjectSelection.gdictionary
         [ddgen] [Info]    Generating datatypes/com/sap/pcuigp/xssutils/pernr/model/types/RHXSS_SER_EMPLOYEE.gdictionary
         [ddgen] [Info]    Generating dbtables/sdmDeployDd.xml
         [ddgen] [Info]    Generation finished (8 seconds)
         [ddgen]
         [timer] Dictionary generator finished in 8.499 seconds
          [echo] Starting WebDynpro generator
         [wdgen]
         [wdgen] [Info]    Property deployment is true: Deployment information is provided!
         [wdgen] [Info]    Property sourcepath: /usr/sap/<SID>/JC51/j2ee/cluster/server0/temp/CBS/27/.B/6448/DCs/sap.com/mss/mol/_comp/src/packages
         [wdgen] [Info]    Property targetpath: /usr/sap/<SID>/JC51/j2ee/cluster/server0/temp/CBS/27/.B/6448/t/64CE6536D7BD962CDCB65DC8A7640047/gen_wdp
         [wdgen] [Info]    Property archivename: sap.commssmol
         [wdgen] [Info]    Property vendor: sap.com
         [wdgen] [Info]    Property dcname: mss/mol
         [wdgen] [Info]    Property projectlanguage: de
         [wdgen] [Info]    Property language: Available languages are automatically determined!
         [wdgen] WARNING: Metadata of component FcLineItemMonitor is not valid! Component "//WebDynpro/Component:com.sap.mss.mol.monitor.FcLineItemMonitor", Role "ComponentController": A minimum of 1 object(s) is required
         [wdgen] [Error]   com.sap.mss.mol.monitor.VcLineItemMonitorView --> TextView rulename_editor [text]: Context element and property are not compatible
         [wdgen] [Error]   com.sap.mss.mol.monitor.VcLineItemMonitorView --> TextView count_editor [text]: Context element and property are not compatible
         [wdgen] [Error]   com.sap.mss.mol.monitor.VcLineItemMonitorView --> TextView evaldate [text]: Context element and property are not compatible
         [wdgen] [Error]   com.sap.mss.mol.monitor.VcLineItemMonitorView --> TextView evaltime_2 [text]: Context element and property are not compatible
         [wdgen] [Info]    com.sap.mss.mol.monitor.VcLineItemMonitorView --> TransparentContainer tcActionBar [Children]: Container does not contain children
         [wdgen] [Error]   com.sap.mss.mol.monitor.VcLineItemMonitorView --> Button btnEditRule [visible]: Context element and property are not compatible
         [wdgen] [Error]   com.sap.mss.mol.monitor.VcLineItemMonitorView --> Button btnCreateRule [visible]: Context element and property are not compatible
         [wdgen] [Error]   com.sap.mss.mol.monitor.VcLineItemMonitorView --> Button btnCopyRule [visible]: Context element and property are not compatible
         [wdgen] [Error]   com.sap.mss.mol.monitor.VcLineItemMonitorView --> Button btnDeleteRule [visible]: Context element and property are not compatible
         [wdgen] [Error]   com.sap.mss.mol.monitor.VcLineItemMonitorView --> Button btnPersonalize [visible]: Context element and property are not compatible
         [wdgen] [Error]   com.sap.mss.mol.monitor.VcLineItemMonitorView --> TextView itemtxt_editor [text]: Context element and property are not compatible
         [wdgen] [Error]   com.sap.mss.mol.monitor.VcLineItemMonitorView --> LinkToURL verakUser_url [text]: Context element and property are not compatible
         [wdgen] [Error]   com.sap.mss.mol.monitor.VcLineItemMonitorView --> LinkToURL verakUser_url [reference]: Context element and property are not compatible
         [wdgen] [Error]   com.sap.mss.mol.monitor.VcLineItemMonitorView --> TextView itemdate_editor [text]: Context element and property are not compatible
         [wdgen] [Error]   com.sap.mss.mol.monitor.VcLineItemMonitorView --> TextView value_editor [text]: Context element and property are not compatible
         [wdgen] [Error]   com.sap.mss.mol.monitor.VcLineItemMonitorView --> TextView waers_editor [text]: Context element and property are not compatible
         [wdgen] [Error]   com.sap.mss.mol.monitor.VcLineItemMonitorView --> Image confirmedImage [tooltip]: Context element and property are not compatible
         [wdgen] [Error]   com.sap.mss.mol.monitor.VcLineItemMonitorView --> Image confirmedImage [alt]: Context element and property are not compatible
         [wdgen] [Error]   com.sap.mss.mol.monitor.VcLineItemMonitorView --> Image confirmedImage [source]: Context element and property are not compatible
         [wdgen] [Error]   com.sap.mss.mol.monitor.VcLineItemMonitorView --> TextView actionTaken_editor [text]: Context element and property are not compatible
         [wdgen] [Error]   com.sap.mss.mol.monitor.VcLineItemMonitorView --> TextView tvCurrency [text]: Context element and property are not compatible
         [wdgen] [Info]    com.sap.mss.mol.monitor.VcLineItemMonitorView --> TransparentContainer DetailContainer [Children]: Container does not contain children
         [wdgen] [Error]   com.sap.mss.mol.monitor.VcLineItemMonitorView --> UIElement UpdateTrigger [enabled]: Context element and property are not compatible
         [wdgen] [Info]    com.sap.mss.mol.monitor.VcLineItemMonitorView --> TextView tvStatus: UIElement does not have a label
         [wdgen] [Info]    com.sap.mss.mol.monitor.VcLineItemMonitorView --> TextView evaldate: UIElement does not have a label
         [wdgen] [Info]    com.sap.mss.mol.monitor.VcLineItemMonitorView --> TextView evaltime_2: UIElement does not have a label
         [wdgen] [Info]    com.sap.mss.mol.monitor.VcLineItemMonitorView --> LinkToAction LinkUpdate: UIElement does not have a label
         [wdgen] [Info]    com.sap.mss.mol.monitor.VcLineItemMonitorView --> TextView tvUpdateRunning: UIElement does not have a label
         [wdgen] [Info]    com.sap.mss.mol.monitor.VcLineItemMonitorView --> Caption RuleGroup_Header: UIElement does not have a label
         [wdgen] [Info]    com.sap.mss.mol.monitor.VcLineItemMonitorView --> TextView tvCurrency: UIElement does not have a label
         [wdgen] [Info]    com.sap.mss.mol.monitor.VcLineItemMonitorView --> Caption Details_Header: UIElement does not have a label
         [wdgen] [Info]    com.sap.mss.mol.monitor.VcLineItemMonitorView --> Caption ItemDetails_Header: UIElement does not have a label
         [wdgen] [Error]   com.sap.mss.mol.monitor.RuleMaintView --> TextView rulename_editor [text]: Context element and property are not compatible
         [wdgen] [Error]   com.sap.mss.mol.monitor.RuleMaintView --> TextView ruleid_editor [text]: Context element and property are not compatible
         [wdgen] [Error]   com.sap.mss.mol.monitor.RuleMaintView --> InputField inDescription [value]: Context element and property are not compatible
         [wdgen] [Error]   com.sap.mss.mol.monitor.RuleMaintView --> CheckBox cbShowAlert [checked]: Context element and property are not compatible
         [wdgen] [Error]   com.sap.mss.mol.monitor.RuleMaintView --> DropDownByKey ddKF1 [selectedKey]: Context element and property are not compatible
         [wdgen] [Error]   com.sap.mss.mol.monitor.RuleMaintView --> DropDownByKey ddSign [selectedKey]: Context element and property are not compatible
         [wdgen] [Error]   com.sap.mss.mol.monitor.RuleMaintView --> InputField inLimitAbs [value]: Context element and property are not compatible
         [wdgen] [Error]   com.sap.mss.mol.monitor.RuleMaintView --> DropDownByKey ddAbsRel [selectedKey]: Context element and property are not compatible
         [wdgen] [Error]   com.sap.mss.mol.monitor.RuleMaintView --> InputField InputTimestamp [value]: Context element and property are not compatible
         [wdgen] [Error]   com.sap.mss.mol.monitor.RuleMaintView --> Tree treeObject [title]: Context element and property are not compatible
         [wdgen] [Error]   com.sap.mss.mol.monitor.RuleMaintView --> Tree treeObject [tooltip]: Context element and property are not compatible
         [wdgen] [Error]   com.sap.mss.mol.monitor.RuleMaintView --> Tree treeKfspecgl [title]: Context element and property are not compatible
         [wdgen] [Error]   com.sap.mss.mol.monitor.RuleMaintView --> Tree treeKfspecgl [tooltip]: Context element and property are not compatible
         [wdgen] [Info]    com.sap.mss.mol.monitor.RuleMaintView --> Caption tableRulesHeader: UIElement does not have a label
         [wdgen] [Info]    com.sap.mss.mol.monitor.RuleMaintView --> CheckBox cbShowAlert: UIElement does not have a label
         [wdgen] [Warning] com.sap.mss.mol.monitor.RuleMaintView --> DropDownByKey ddSign: UIElement does not have a label
         [wdgen] [Warning] com.sap.mss.mol.monitor.RuleMaintView --> InputField inLimitAbs: UIElement does not have a label
         [wdgen] [Warning] com.sap.mss.mol.monitor.RuleMaintView --> DropDownByKey ddAbsRel: UIElement does not have a label
         [wdgen] [Info]    Catching throwable null
         [wdgen] [Info]    com.sap.webdynpro.generation.ant.GenerationAntTaskError
         [wdgen]      at com.sap.webdynpro.generation.ant.GenerationAnt.showCheckResult(GenerationAnt.java:157)
         [wdgen]      at com.sap.tc.buildcontroller.CBSBuildController.mainLoop(CBSBuildController.java:207)
         [wdgen]      at com.sap.tc.buildcontroller.CBSBuildController.main(CBSBuildController.java:168)
         [wdgen] ERROR: Unknown exception during generation null (com.sap.webdynpro.generation.ant.GenerationAntTaskError)
         [wdgen] ERROR: Generation failed due to errors (7 seconds)
    Error: /usr/sap/<SID>/JC51/j2ee/cluster/server0/temp/CBS/27/.B/6448/DCs/sap.com/mss/mol/_comp/gen/default/logs/build.xml:99: [Error]   Generation failed!
         at com.sap.webdynpro.generation.ant.WDGenAntTask.execute(WDGenAntTask.java:254)
         at org.apache.tools.ant.UnknownElement.execute(UnknownElement.java:275)
         at org.apache.tools.ant.Task.perform(Task.java:364)
         at com.sap.tc.buildcontroller.CBSBuildController.mainLoop(CBSBuildController.java:207)
         at com.sap.tc.buildcontroller.CBSBuildController.main(CBSBuildController.java:168)
    Ant runtime 16.597 seconds
    Ant build finished with ERRORS
    [Error]   Generation failed!
    Error: Build stopped due to an error: [Error]   Generation failed!
    Build plugin finished at 2008-11-28 22:04:32 GMT-07:00 (MST)
    Total build plugin runtime: 20.648 seconds
    Build finished with ERROR
    What can be done to resolve these error and trigger the build again.
    Bhavin

    Resloved with SAP help.
    1. Ran configuration wizard for CBS, CMS & DTR
    2. Changes few JVM parameters (increased heap size, jdbc pool size)

Maybe you are looking for

  • Setting JFrame Color

    Hi everyone I have a problem setting JFrame background color and the Foreground color. Here is part of my code where I have the problem. import java.io.*; import java.awt.*; import java.awt.event.*; import java.text.*; public class userInfo extends F

  • What's the main difference of different Type

    I always see Type 3 or Typr 4 jdbc driver. What's the main difference between them? Raymond

  • OSB 10.3: Enabling JConsole

    Hi, I am trying to enable JConsole on OSB (Weblogic AdminServer/JDK 1.6.0_05). Used the following parameters to enable JMX: -Dcom.sun.management.jmxremote -Dcom.sun.management.jmxremote.port=8889 -Dcom.sun.management.jmxremote.local.only=false -Dcom.

  • Bapi calling webservice

    HI friends, I am new in ABAP world, i wonder can i call a webservice from my BAPI ? thanks Pranav

  • Backing up .epub files from iBooks

    Hi, I wonder if there is a way to back up your iBooks .epub files? They're not purchased eBooks, rather, they're fanfiction downloaded from fanwork archive like http://archiveofourown.org/ (if you check the works, there's a neat little link where you