Statistic on Views.

Hi All,
I was wondering if there are way to get statistics on bex query views. (eg last run / last modied) 
Our intention is to identify view usage rather than the query.
Regards

Dear Mark,
   seen this?
http://help.sap.com/saphelp_nw04s/helpdata/en/b0/5a213c07fdb66be10000000a11405a/content.htm
Thanks,
Raj

Similar Messages

  • How to know report log (statistic of viewed report)

    Hi All,
    I wanna know how to get viewed report log (statistic)
    The statistic contains the name of query report, frequency using report based on user name.
    Thank you,
    Jane

    Hi...for all of statistics you can find some infocubes and reports...you can read the documentation here...
    http://help.sap.com/saphelp_nw70/helpdata/en/46/f9bd5b0d40537de10000000a1553f6/frameset.htm
    Regards

  • Linear Regression - How to view t-statistic and p-value

    How can we view the t-statistic and p-value for the predictors in a linear regression or multiple linear regression model?

    The R code sample for LM would be as follows: model <- lm(dependent_var ~ independent_var, data=data.set)
    In the Azure ML gallery there are a few samples which can help you get set up with the experiment. (Sample name: Exploring the Usage of R in AML)

  • Essbase tuning advisor / statistic viewer?

    Hi,
    Is there any tool that can help me fine tune my calculation scripts / business rules? I would like to know for instance, how many blocks are being affected with a couple of fixes, or the exact members where a fix calculation will have effect if I used a function like @Ancestors or @Children inside a fix.
    It would be mostly to make some tests and check the time they take calculating.
    Thank you

    I would like to know for instance, how many blocks are being affected with a couple of fixes^^^SET MSG SUMMARY or SET MSG DETAIL and the application log will tell you that. It's sort of a pain to work with, but it's all there.
    Regards,
    Cameron Lackpour

  • Please help me open an iMovie trailer, moved to my iPad from my PC. I can view it on my iPad but am unable to import it into the app to edit it.

    I created a trailer on an iPad at a workshop to use in my classroom. I retrieved it from my PC and can view it on my iPad mini, but the iMovie app won't open it. I need to edit it. Desperately need help. I want to show it to my students tomorrow. Thanks! I've checked other threads here but can't find a solution.

    Hi Leica 42,
    I never said that a PP file moved to your ipad cannot be straight edited in keynote. Nor did I say that you needed a third party solution...
    My response was to offer an option (Jump) because like many others, you're having problems.
    Check out the app store and user Keynote reviews. There are more (by 1) 1 star ratings than 5 star ratings for the current version. There are 72 ratings of 3 or more stars while there are 78 ratings of one or two stars. In terms of statistical significance, the numbers are pretty equal, but I prefer to deal with an app that shows more top stars and fewer bottom stars. Some posters have been on here with your same question, so I'm thinking maybe, just perhaps, Keynote is not perfect. 
    Pro tech reviews vary, many given at the time it was released. Many do not point out that they actually used Keynote before the review. Google Keynote for more pro tech reviews where reviewers have actually used the app.
    http://www.techradar.com/us/reviews/pc-mac/software/business-and-finance-softwar e/apple-keynote-ipad-689424/review
    http://www.theverge.com/2013/10/29/5042880/apple-iwork-2013-refresh-complaints
    http://www.igeeksblog.com/ipad-presentation-apps/

  • No Data in Field Scrap quantity (AMGLG) after Append DataSource 2LIS_04_P_COMP (Component View from PP/PP-PI)

    Hi experts,
    we enhanced the BI Content standard DataSource 2LIS_04_P_COMP "Component View from PP/PP-PI" with the field AMGLG (Scrap quantity) per maintenance of the relevant extract structure MC04P_0COM via LBWE:
    The enhancement of the extract structure worked fine.
    Unfortunately the enhanced field is not filled after the Statistical Setup of the LIS Application 04 "Production" (t/code OLI4BW).
    RSA3 -> 2LIS_04_P_COMP:
    Any ideas or experiences?
    Many thanks in advance and regards,
    Michael

    Attached one example:
    BW DataSource 2LIS_04_P_COMP:
    ERP Summarized BOM Multilevel (t/code CS13) Component 456:
    ERP Summarized BOM Multilevel (t/code CS13) Component 301:
    Any ideas why there are no values for the field AMGLG extracted?
    Thanks and regards,
    Michael

  • Using EL to get values from View Objects

    Hello again!
    I' m using jdev 11.1.1.4.0, with adf and business components
    Is there any way i can get a value to an af:outputText from a view object's field without
    a) having to expose all fields that I need in bindings tab?
    b) having to declare all fields to a backing bean
    I have a statistical table with about 30 numbers in a record and it would make my project impossible to follow!
    Some EL on the af:outputText with a parameter to a function on my backing bean, would sound perfect, but I've read that it's impossible...
    Any clues?
    Thank you for your efforts!
    Nikos

    I did not understand you.
    will the drag and drop your attribute on your page will solve your problem?
    or do you need to create attributeValues biding for single attribute? if yes do the following:
    1- From your page, right click and select Go to Page Definition.
    2- From the binding section, click the plus green icon to create a new control biding, and choose attributeValues from the list.
    3- select your data source from the list or create a new one, then select the attribute you need.
    4- now you have a biding, and you can set the value for your output text.
    value =#{biding.yourAttribute.inputValue}

  • Creating query on Bex - Quaterly comparison for statistical & Actual

    Hi All,
    I would like to create a query for 'Quarterly comparison for statistical & Actual periods'.
    My Key Figures should be
    1) Plan 1st Qtr (Fiscal year, Period: 1 to 3, Value type : 1(Plan), Version : 0(Plan/actual), Valuation View: actual Value).
    2)1st Qtr (Fiscal year, Period: 1 to 3, "Value type : 4(Actual),11(Actual statistical)", Version : 0(Plan/actual), Valuation View: actual Value).
    3)Var 1st Qt (Plan 1st qtr - 1st Qtr)
    same thing for 4 Quaters. finally with
    4)Plan Year (Fiscal year, Period: 1 to 12, Value type : 1(Plan), Version : 0(Plan/actual), Valuation View: actual Value).
    I created a structure and created key figures with selections and formulas as required. But I did not see any data when I ran this query.
    The report was generated with 'no applicable data'.
    I need to create this query with plan 1st Qtr, Ist Qtr, Var 1st Qtr, Plan 2nd Qtr, 2nd Qtr, Var 2nd Qtr, Plan 3rd Qtr, 3rd Qtr, Var 3rd Qtr, Plan 4th Qtr, 4th Qtr, Var 4th Qtr, Plan year. key figures.
    Please let me know how can I create this query with these Key Figiures.
    Any help would be appreciated. Please respond with the reply.
    Thanks,
    Aparna.

    Hi
    The best way is then to run a report with your KF without any restriction, and the different chars in the drill down: Fiscal year, Period:, Value type,  Version , Valuation View
    Then you can check that you have some information with the combination of values of your chars:
    Fiscal year, Period: 1 to 3, Value type : 1(Plan), Version : 0(Plan/actual), Valuation View: actual Value.
    If you find a actual Value in the fiscal period you are looking at, for the period 1 to 3, for the Valuation type 1, for the version 0, then create arestricted KF by adding the restrictions one at a time....You moght discover why you do not get the results
    PY

  • How do I use Dreamweaver CC 2014.1 Live View to copy complex formatted text into web page?

    Since Design View has been removed from Fluid Grid Layouts, and could be removed completely in next upgrade, I cannot use Live View to place complex text into my web pages.
    Our business requires the listing of complex job descriptions on our web pages.  Design View worked fine, Live View does not. Both examples are noted below.
    Using Design View the web page rendered this way:
    Research & Analytics Director
    Position# 2469
    Location:  Midwest
    Education Requirements:  MS degree in Mathematics, Statistics, Economics, Engineering, Actuarial Science or related field or equivalent designation, such as FCAS.
    Other Requirements:  Include:
    Job Competencies:
    Achieve Results.
    Be Accountable.
    Lead Change.
    Lead Corp Vision & Strategy.
    Lead People.
    Maximize Customer Experience.
    Specialized Knowledge and Skills Requirements:
    Demonstrated ability to develop strategic partnerships.
    Demonstrated ability to identify potential issues, and to proactively work to the mitigation of those issues.
    Demonstrated experience communicating business implications of complex data relationships and results of statistical models to multiple business partners.
    Demonstrated experience formulating, approaching, and solving problems in massive, complex datasets.
    Demonstrated experience in public speaking and use of appropriate presentation skills.
    Demonstrated experience interfacing with business clients and driving solution discussions with both IT and business stakeholders.
    Demonstrated experience performing advanced statistical analysis, including generalized linear models, decision trees, neural networks, etc., to discover business insights and develop predictive models.
    Demonstrated experience with statistical and modeling software tools, such as SAS or R.
    Demonstrated management or leadership experience.
    Solid knowledge and understanding of mathematical modeling and research.
    Salary Range:  $146,300 to $243,800+
    Description:  This position is responsible for oversight and strategic direction of advanced analytics, including predictive modeling, that drive business performance consistent with company goals and objectives.  Works with internal business partners to understand and validate scope of advanced analytics projects and directs projects teams to support those projects.  Collaborates within the division and cross-divisionally to develop and integrate statistical models into various processes. Oversees research of new and innovative analytic tools and techniques.  Monitors industry trends in analytics and investigates new concepts, ideas, and data sources.  Primary accountabilities include:
    Advanced Analytics Oversight (40%):
    Provides oversight and direction for the design, development, and evaluation of predictive models and advanced algorithms that lead to business solutions.
    Directs the most complex and vital analytics work critical to the organization.
    Oversees advanced exploratory analytics that produce a variety of business solutions.
    Conducts peer review on technical aspects of projects.
    Partners with business areas (Commercial Lab, EDM, IS, etc.) on data innovation.
    Portfolio Strategy and Management (25%):
    Collaborates with Strategic Data & Analytics Vice President as well as business partners internal and external to the division to develop and execute advanced analytics strategy and project portfolio.
    Leverages business acumen and domain expertise in directing advanced analytics strategy and application.
    Aligns plans to divisional and corporate objectives/goals and integrates within the corporate planning cycle.
    Monitors and analyzes advanced analytics resources. Assesses needs and appropriately allocates resources to priorities and initiatives.
    Works with divisional management to improve work processes that impact work environment and divisional resources.
    Provides overall portfolio/project management oversight and direction for a variety of advanced analytics project.  Oversees measurement, monitoring and reporting of project progress and resource utilization.  Manages project resources to ensure timely delivery of projects consistent with divisional goals and objectives.
    Coordinates with business partners (Analytics Strategy, etc.) on knowledge management and participates in business area meetings. Maintains holistic view of the business.  Ensures rapid delivery and execution of insights derived from data analytics into the organization.
    Analytic Tools and Techniques Research Oversight (15%):
    Sets the vision for the use of new and innovative tools and technology.
    Maintains and fosters an industry awareness of new developments in analytics techniques and tools, and ensures quick execution in their use within the department.
    Interfaces with Enterprise Data Management (EDM) and Information Services (IS) on evolving technological and data needs and requirements.
    Strategies Linked to the Division's Business Goals/Results (10%):
    Establishes, communicates, and implements departmental plans, objectives, and strategies.
    Participates as a member of the Management Team.
    Maintains an active awareness of our Client's business environments, corporate culture, and structure to support key decision-making.
    Management/Leadership for Department or Unit (10%):
    Manages direct reports, systems, and projects to achieve department/unit goals in accordance with Company policies and practices.
    Prepares and analyzes department/unit plans and reports.
    Provides leadership by exhibiting influence and expertise, thus affecting the results of the operating area.
    Creates an effective work environment by developing a common vision, setting clear objectives, expecting teamwork, recognizing outstanding performance, and maintaining open communications.
    Develops staff through coaching, providing performance feedback, providing effective performance assessments, and establishing performance & development plans.
    Using Live View the web page rendered this way:
    Research & Analytics Director Position# 2469 Location: Midwest Education Requirements: MS degree in Mathematics, Statistics, Economics, Engineering, Actuarial Science or related field or equivalent designation, such as FCAS. Other Requirements: Include: • Job Competencies: o Achieve Results. o Be Accountable. o Lead Change. o Lead Corp Vision & Strategy. o Lead People. o Maximize Customer Experience. • Specialized Knowledge and Skills Requirements: o Demonstrated ability to develop strategic partnerships. o Demonstrated ability to identify potential issues, and to proactively work to the mitigation of those issues. o Demonstrated experience communicating business implications of complex data relationships and results of statistical models to multiple business partners. o Demonstrated experience formulating, approaching, and solving problems in massive, complex datasets. o Demonstrated experience in public speaking and use of appropriate presentation skills. o Demonstrated experience interfacing with business clients and driving solution discussions with both IT and business stakeholders. o Demonstrated experience performing advanced statistical analysis, including generalized linear models, decision trees, neural networks, etc., to discover business insights and develop predictive models. o Demonstrated experience with statistical and modeling software tools, such as SAS or R. o Demonstrated management or leadership experience. o Solid knowledge and understanding of mathematical modeling and research. Salary Range: $146,300 to $243,800+ Description: This position is responsible for oversight and strategic direction of advanced analytics, including predictive modeling, that drive business performance consistent with company goals and objectives. Works with internal business partners to understand and validate scope of advanced analytics projects and directs projects teams to support those projects. Collaborates within the division and cross-divisionally to develop and integrate statistical models into various processes. Oversees research of new and innovative analytic tools and techniques. Monitors industry trends in analytics and investigates new concepts, ideas, and data sources. Primary accountabilities include: • Advanced Analytics Oversight (40%): o Provides oversight and direction for the design, development, and evaluation of predictive models and advanced algorithms that lead to business solutions. o Directs the most complex and vital analytics work critical to the organization. o Oversees advanced exploratory analytics that produce a variety of business solutions. o Conducts peer review on technical aspects of projects. o Partners with business areas (Commercial Lab, EDM, IS, etc.) on data innovation. • Portfolio Strategy and Management (25%): o Collaborates with Strategic Data & Analytics Vice President as well as business partners internal and external to the division to develop and execute advanced analytics strategy and project portfolio. o Leverages business acumen and domain expertise in directing advanced analytics strategy and application. o Aligns plans to divisional and corporate objectives/goals and integrates within the corporate planning cycle. o Monitors and analyzes advanced analytics resources. Assesses needs and appropriately allocates resources to priorities and initiatives. o Works with divisional management to improve work processes that impact work environment and divisional resources. o Provides overall portfolio/project management oversight and direction for a variety of advanced analytics project. Oversees measurement, monitoring and reporting of project progress and resource utilization. Manages project resources to ensure timely delivery of projects consistent with divisional goals and objectives. o Coordinates with business partners (Analytics Strategy, etc.) on knowledge management and participates in business area meetings. Maintains holistic view of the business. Ensures rapid delivery and execution of insights derived from data analytics into the organization. • Analytic Tools and Techniques Research Oversight (15%): o Sets the vision for the use of new and innovative tools and technology. o Maintains and fosters an industry awareness of new developments in analytics techniques and tools, and ensures quick execution in their use within the department. o Interfaces with Enterprise Data Management (EDM) and Information Services (IS) on evolving technological and data needs and requirements. • Strategies Linked to the Division's Business Goals/Results (10%): o Establishes, communicates, and implements departmental plans, objectives, and strategies. o Participates as a member of the Management Team. o Maintains an active awareness of our Client's business environments, corporate culture, and structure to support key decision-making. • Management/Leadership for Department or Unit (10%): o Manages direct reports, systems, and projects to achieve department/unit goals in accordance with Company policies and practices. o Prepares and analyzes department/unit plans and reports. o Provides leadership by exhibiting influence and expertise, thus affecting the results of the operating area. o Creates an effective work environment by developing a common vision, setting clear objectives, expecting teamwork, recognizing outstanding performance, and maintaining open communications. o Develops staff through coaching, providing performance feedback, providing effective performance assessments, and establishing performance & development plans.
    Obviously, as seen above, I'm doing something wrong or Live View does not do all the functions that Design View use to do.  I realize I can still access Design View in non-Fluid Grid Layout documents and that is my current work-around.  I copy from my Word file into Design View available in non-Fluid, then copy that code into my Fluid Grid Layout document, not real efficient use of my time.  And, even that work-around may not be available after the next upgrade.

    Hans thank you for your reply and attention to this matter.  Moreover, with your steeped expertise I do not, in any fashion, intend to be flippant or capricious. 
    However, I do not believe the reply is on point since I'm not discussing the insertion of an HTML tag/element; but rather, the placing of complex formatted text into my various web pages.
    Design View and  2014.1 Live View seem to serve different functions.  That is, Design View, for us, serves as a highly sophisticated HTML5 code generator that allows us to properly display complex formatted text that originally is created as a Word docx file.  As I noted in my original post, the complex formatted text from the Word file is then pasted into Design View and thus properly renders on the web page.  I invite you to try this yourself.  Simply copy my properly formatted text in my original post (that can be found at "Using Design View the web page rendered this way:") to a Word file, and it will retain the complex formatting, then copy that Word file text to both Design View and  2014.1 Live View files.  Please apprise me of your results.
    Discussion has been had in at least two separate threads on the Design View disappearance in Fluid Grid.  In https://forums.adobe.com/thread/1597260 staff member Lalita wrote "It would be helpful for us if you list the issues you are seeing with fluid grid live view editing."  And in https://forums.adobe.com/message/6807088#6807088 staff member Subhadeep wrote "If you can list down the workflows you are trying while editing in Live View & the issues you are facing in doing so, it will help us understand what is amiss & suggest alternative workflows to do the same."  I believe I've pointed out issues when attempting to use Live View for functions or work flow that customarily had been done in Design View.
    Being a CC subscriber, if someone can provide me with a workflow using other additional Adobe products to get my complex formatted text from Word ultimately to my Dreamweaver web page, be it an image insert or whatever, I would take the suggestions.
    The difference in the coding of my example in my original post is as follows.
    Code when using Design View is:
    <p><strong>Research &amp; Analytics Director</strong><br>
      <strong>Position# 2469</strong><br>
      <strong>Location:</strong>  Midwest<br>
      <strong>Education Requirements:</strong>  MS degree in Mathematics, Statistics,  Economics, Engineering, Actuarial Science or related field or equivalent  designation, such as FCAS.<br>
      <strong>Other Requirements:</strong>  Include:</p>
    <ul type="disc">
      <li>Job       Competencies:</li>
      <ul type="circle">
        <li>Achieve        Results.</li>
        <li>Be        Accountable.</li>
        <li>Lead        Change.</li>
        <li>Lead Corp        Vision &amp; Strategy.</li>
        <li>Lead        People.</li>
        <li>Maximize        Customer Experience.</li>
      </ul>
      <li>Specialized       Knowledge and Skills Requirements:</li>
      <ul type="circle">
        <li>Demonstrated        ability to develop strategic partnerships.</li>
        <li>Demonstrated        ability to identify potential issues, and to proactively work to the        mitigation of those issues.</li>
        <li>Demonstrated        experience communicating business implications of complex data        relationships and results of statistical models to multiple business        partners.</li>
        <li>Demonstrated        experience formulating, approaching, and solving problems in massive,        complex datasets.</li>
        <li>Demonstrated        experience in public speaking and use of appropriate presentation skills.</li>
        <li>Demonstrated        experience interfacing with business clients and driving solution        discussions with both IT and business stakeholders.</li>
        <li>Demonstrated        experience performing advanced statistical analysis, including        generalized linear models, decision trees, neural networks, etc., to        discover business insights and develop predictive models.</li>
        <li>Demonstrated        experience with statistical and modeling software tools, such as SAS or        R.</li>
        <li>Demonstrated        management or leadership experience.</li>
        <li>Solid        knowledge and understanding of mathematical modeling and research.</li>
      </ul>
    </ul>
    <p><strong>Salary Range:</strong>  $146,300 to  $243,800+<br>
      <strong>Description:</strong>  This position is  responsible for oversight and strategic direction of advanced analytics,  including predictive modeling, that drive business performance consistent with  company goals and objectives.  Works with internal business partners to  understand and validate scope of advanced analytics projects and directs  projects teams to support those projects.  Collaborates within the  division and cross-divisionally to develop and integrate statistical models into  various processes. Oversees research of new and innovative analytic tools and  techniques.  Monitors industry trends in analytics and investigates new  concepts, ideas, and data sources.  Primary accountabilities include:</p>
    <ul type="disc">
      <li>Advanced       Analytics Oversight (40%):</li>
      <ul type="circle">
        <li>Provides        oversight and direction for the design, development, and evaluation of        predictive models and advanced algorithms that lead to business        solutions.</li>
        <li>Directs        the most complex and vital analytics work critical to the organization.</li>
        <li>Oversees        advanced exploratory analytics that produce a variety of business        solutions.</li>
        <li>Conducts        peer review on technical aspects of projects.</li>
        <li>Partners        with business areas (Commercial Lab, EDM, IS, etc.) on data innovation.</li>
      </ul>
      <li>Portfolio       Strategy and Management (25%):</li>
      <ul type="circle">
        <li>Collaborates        with Strategic Data &amp; Analytics Vice President as well as business        partners internal and external to the division to develop and execute        advanced analytics strategy and project portfolio.</li>
        <li>Leverages        business acumen and domain expertise in directing advanced analytics        strategy and application.</li>
        <li>Aligns        plans to divisional and corporate objectives/goals and integrates within        the corporate planning cycle.</li>
        <li>Monitors        and analyzes advanced analytics resources. Assesses needs and        appropriately allocates resources to priorities and initiatives.</li>
        <li>Works        with divisional management to improve work processes that impact work        environment and divisional resources.</li>
        <li>Provides        overall portfolio/project management oversight and direction for a        variety of advanced analytics project.  Oversees measurement,        monitoring and reporting of project progress and resource        utilization.  Manages project resources to ensure timely delivery of        projects consistent with divisional goals and objectives.</li>
        <li>Coordinates        with business partners (Analytics Strategy, etc.) on knowledge management        and participates in business area meetings. Maintains holistic view of        the business.  Ensures rapid delivery and execution of insights        derived from data analytics into the organization.</li>
      </ul>
      <li>Analytic       Tools and Techniques Research Oversight (15%):</li>
      <ul type="circle">
        <li>Sets the        vision for the use of new and innovative tools and technology.</li>
        <li>Maintains        and fosters an industry awareness of new developments in analytics        techniques and tools, and ensures quick execution in their use within the        department.</li>
        <li>Interfaces        with Enterprise Data Management (EDM) and Information Services (IS) on        evolving technological and data needs and requirements.</li>
      </ul>
      <li>Strategies       Linked to the Division's Business Goals/Results (10%):</li>
      <ul type="circle">
        <li>Establishes,        communicates, and implements departmental plans, objectives, and        strategies.</li>
        <li>Participates        as a member of the Management Team.</li>
        <li>Maintains        an active awareness of our Client's business environments, corporate        culture, and structure to support key decision-making.</li>
      </ul>
      <li>Management/Leadership       for Department or Unit (10%):</li>
      <ul type="circle">
        <li>Manages        direct reports, systems, and projects to achieve department/unit goals in        accordance with Company policies and practices.</li>
        <li>Prepares        and analyzes department/unit plans and reports.</li>
        <li>Provides        leadership by exhibiting influence and expertise, thus affecting the        results of the operating area.</li>
        <li>Creates        an effective work environment by developing a common vision, setting        clear objectives, expecting teamwork, recognizing outstanding performance,        and maintaining open communications.</li>
        <li>Develops        staff through coaching, providing performance feedback, providing        effective performance assessments, and establishing performance &amp;        development plans.</li>
      </ul>
    </ul>
    Code when using Live View is:
      <div id="liveview" class="fluid">Research &amp; Analytics Director Position# 2469 Location: Midwest Education Requirements: MS degree in Mathematics, Statistics, Economics, Engineering, Actuarial Science or related field or equivalent designation, such as FCAS. Other Requirements: Include: • Job Competencies: o Achieve Results. o Be Accountable. o Lead Change. o Lead Corp Vision &amp; Strategy. o Lead People. o Maximize Customer Experience. • Specialized Knowledge and Skills Requirements: o Demonstrated ability to develop strategic partnerships. o Demonstrated ability to identify potential issues, and to proactively work to the mitigation of those issues. o Demonstrated experience communicating business implications of complex data relationships and results of statistical models to multiple business partners. o Demonstrated experience formulating, approaching, and solving problems in massive, complex datasets. o Demonstrated experience in public speaking and use of appropriate presentation skills. o Demonstrated experience interfacing with business clients and driving solution discussions with both IT and business stakeholders. o Demonstrated experience performing advanced statistical analysis, including generalized linear models, decision trees, neural networks, etc., to discover business insights and develop predictive models. o Demonstrated experience with statistical and modeling software tools, such as SAS or R. o Demonstrated management or leadership experience. o Solid knowledge and understanding of mathematical modeling and research. Salary Range: $146,300 to $243,800+ Description: This position is responsible for oversight and strategic direction of advanced analytics, including predictive modeling, that drive business performance consistent with company goals and objectives. Works with internal business partners to understand and validate scope of advanced analytics projects and directs projects teams to support those projects. Collaborates within the division and cross-divisionally to develop and integrate statistical models into various processes. Oversees research of new and innovative analytic tools and techniques. Monitors industry trends in analytics and investigates new concepts, ideas, and data sources. Primary accountabilities include: • Advanced Analytics Oversight (40%): o Provides oversight and direction for the design, development, and evaluation of predictive models and advanced algorithms that lead to business solutions. o Directs the most complex and vital analytics work critical to the organization. o Oversees advanced exploratory analytics that produce a variety of business solutions. o Conducts peer review on technical aspects of projects. o Partners with business areas (Commercial Lab, EDM, IS, etc.) on data innovation. • Portfolio Strategy and Management (25%): o Collaborates with Strategic Data &amp; Analytics Vice President as well as business partners internal and external to the division to develop and execute advanced analytics strategy and project portfolio. o Leverages business acumen and domain expertise in directing advanced analytics strategy and application. o Aligns plans to divisional and corporate objectives/goals and integrates within the corporate planning cycle. o Monitors and analyzes advanced analytics resources. Assesses needs and appropriately allocates resources to priorities and initiatives. o Works with divisional management to improve work processes that impact work environment and divisional resources. o Provides overall portfolio/project management oversight and direction for a variety of advanced analytics project. Oversees measurement, monitoring and reporting of project progress and resource utilization. Manages project resources to ensure timely delivery of projects consistent with divisional goals and objectives. o Coordinates with business partners (Analytics Strategy, etc.) on knowledge management and participates in business area meetings. Maintains holistic view of the business. Ensures rapid delivery and execution of insights derived from data analytics into the organization. • Analytic Tools and Techniques Research Oversight (15%): o Sets the vision for the use of new and innovative tools and technology. o Maintains and fosters an industry awareness of new developments in analytics techniques and tools, and ensures quick execution in their use within the department. o Interfaces with Enterprise Data Management (EDM) and Information Services (IS) on evolving technological and data needs and requirements. • Strategies Linked to the Division's Business Goals/Results (10%): o Establishes, communicates, and implements departmental plans, objectives, and strategies. o Participates as a member of the Management Team. o Maintains an active awareness of our Client's business environments, corporate culture, and structure to support key decision-making. • Management/Leadership for Department or Unit (10%): o Manages direct reports, systems, and projects to achieve department/unit goals in accordance with Company policies and practices. o Prepares and analyzes department/unit plans and reports. o Provides leadership by exhibiting influence and expertise, thus affecting the results of the operating area. o Creates an effective work environment by developing a common vision, setting clear objectives, expecting teamwork, recognizing outstanding performance, and maintaining open communications. o Develops staff through coaching, providing performance feedback, providing effective performance assessments, and establishing performance &amp; development plans.  </div>
    It is readily apparent that Design View is a much needed code generator in the Dreamweaver product.  Live View serves completely other, and needed, purposes.  Therefore, I still, after all the forum discussions, do not understand why Design View has been removed from Fluid Grid.
    Again Hans thank you for your reply.  But unless I'm missing something, citation to the links you provided does not solve my problem.  And that problem is this: If Design View is on its way out and thus will be completely removed from Dreamweaver at some point, then how do I provide my information (i.e. complex executive job descriptions) on my website?
    Apparently a vote is taking place https://forums.adobe.com/ideas/3922 on the issue of placing Design View back into Fluid Grid.  I do not understand why the need for two rather disjunctive functionalities are at the mercy of the democratic process.  These are matters of engineering principles not "town hall gatherings".

  • [oracle 10.2.0.4] My view is not merged

    Hi,
    I have a view which is not merged by the CBO. I mean the CBO decides to apply the filter predicate after the execution of the view.
    Here is the definition of the view
    CREATE OR REPLACE VIEW VUNSCP AS
    SELECT X.DASFM,X.COINT,X.NUCPT,X.RGCOD,X.RGCID,X.CODEV,X.CTDEV,X.CDVRF,X.TXCHJ,X.MTNLV,X.MTVDP,
           LEAD(X.MTNLV+X.MTVDP,1) OVER (PARTITION BY X.COINT,X.NUCPT,X.CDVRF,X.CTDEV,X.CODEV,X.RGCOD,X.RGCID  ORDER BY X.DASFM DESC),
           SUM(X.MTNLV) OVER (PARTITION BY X.COINT,X.CODEV,X.RGCOD)
    FROM SFMCPT XThe query is:
    explain plan for
    select * from VUNSCP where dasfm='30-apr-10';
    select * from table(dbms_xplan.display);
    Plan hash value: 2545326530
    | Id  | Operation           | Name   | Rows  | Bytes |TempSpc| Cost (%CPU)| Time     |
    |   0 | SELECT STATEMENT    |        |    13M|  1529M|       |   195K  (1)| 00:39:11 |
    |*  1 |  VIEW               | VUNSCP |    13M|  1529M|       |   195K  (1)| 00:39:11 |
    |   2 |   WINDOW SORT       |        |    13M|   646M|  1996M|   195K  (1)| 00:39:11 |
    |   3 |    TABLE ACCESS FULL| SFMCPT |    13M|   646M|       | 27991   (4)| 00:05:36 |
    Predicate Information (identified by operation id):
       1 - filter("DASFM"='30-apr-10')You can see that a FTS is performed on SFMCPT (>1 million of rows) and that the filter predicate is applied only after the view has been instantiated.
    So the index on DASFM can't be used.
    This query is returning about 30 000 rows. We see on the plan that the CBO is mistaken beacause it reckons that there's going to be 13M of rows.
    If I add the filter predicate directly on the view'script I get the correct plan:
    explain plan for
    SELECT X.DASFM,X.COINT,X.NUCPT,X.RGCOD,X.RGCID,X.CODEV,X.CTDEV,X.CDVRF,X.TXCHJ,X.MTNLV,X.MTVDP,
           LEAD(X.MTNLV+X.MTVDP,1) OVER (PARTITION BY X.COINT,X.NUCPT,X.CDVRF,X.CTDEV,X.CODEV,X.RGCOD,X.RGCID ORDER BY X.DASFM DESC),
           SUM(X.MTNLV) OVER (PARTITION BY X.COINT,X.CODEV,X.RGCOD)
    FROM SFMCPT X where dasfm='30-apr-10';
    select * from table(dbms_xplan.display);
    Plan hash value: 1865390099
    | Id  | Operation                    | Name    | Rows  | Bytes | Cost (%CPU)| Time     |
    |   0 | SELECT STATEMENT             |         | 14357 |   729K| 13271   (1)| 00:02:40 |
    |   1 |  WINDOW SORT                 |         | 14357 |   729K| 13271   (1)| 00:02:40 |
    |   2 |   TABLE ACCESS BY INDEX ROWID| SFMCPT  | 14357 |   729K| 13269   (1)| 00:02:40 |
    |*  3 |    INDEX RANGE SCAN          | SFMCPT1 | 14357 |       |    67   (0)| 00:00:01 |
    Predicate Information (identified by operation id):
       3 - access("DASFM"='30-apr-10')The index is now used and the rows estimated seem closer to the actual rows.
    I tried several things:
    - disabling the "OPTIMZER_COST_BASED_TRANSFORMATION" hidden parameter
    - use the MERGE hint
    - alter session set optimizer_features_enable = '9.2.0.8';
    All these workarounds don't work => I'm still getting the bad execution plan.
    According to Jonathan LEWIS' s book the 9i optimzer always merge views But here even if I set the optimizer_features_enable parameter to 9i the view is not merged.
    It's sure that the issue is due to the analytical functions but why ?
    Can please someone help me to understand what is going on ?
    Edited by: Ahmed AANGOUR on 5 mai 2010 08:41

    here is the 10053 trace file:
    /oracle/app/oracle/admin/UBIXPROD/udump/ubixprod_ora_24971_10053_optimizer_trace.trc
    Oracle Database 10g Enterprise Edition Release 10.2.0.4.0 - 64bit Production
    With the Partitioning, OLAP, Data Mining and Real Application Testing options
    ORACLE_HOME = /oracle/app/oracle/10.2.0
    System name:    Linux
    Node name:      lin-ubi-test1.ubitrade.com
    Release:        2.6.9-78.0.1.ELsmp
    Version:        #1 SMP Tue Jul 22 18:01:05 EDT 2008
    Machine:        x86_64
    Instance name: UBIXPROD
    Redo thread mounted by this instance: 1
    Oracle process number: 26
    *** 2010-05-04 12:14:51.450
    *** ACTION NAME:() 2010-05-04 12:14:51.450
    *** MODULE NAME:([email protected] (TNS V1-V3)) 2010-05-04 12:14:51.450
    *** SERVICE NAME:(SYS$USERS) 2010-05-04 12:14:51.450
    *** SESSION ID:(135.1512) 2010-05-04 12:14:51.450
    Registered qb: SEL$1 0xa9e139a8 (PARSER)
      signature (): qb_name=SEL$1 nbfros=1 flg=0
        fro(0): flg=5 objn=297481 hint_alias="VUNSCP"@"SEL$1"
    Registered qb: SEL$2 0xa9e0bdd0 (PARSER)
      signature (): qb_name=SEL$2 nbfros=1 flg=0
        fro(0): flg=4 objn=265023 hint_alias="X"@"SEL$2"
    Predicate Move-Around (PM)
    PM: Considering predicate move-around in SEL$1 (#0).
    PM:   Checking validity of predicate move-around in SEL$1 (#0).
    CBQT: Validity checks failed for 3xakq94fcx4td.
    CVM: Considering view merge in query block SEL$1 (#0)
    CVM:   Checking validity of merging SEL$2 (#0)
    CVM: Considering view merge in query block SEL$2 (#0)
    CVM:     CVM bypassed: Window functions in this view
    CBQT: Validity checks failed for 3xakq94fcx4td.
    Subquery Unnest
    SU: Considering subquery unnesting in query block SEL$1 (#0)
    Set-Join Conversion (SJC)
    SJC: Considering set-join conversion in SEL$1 (#0).
    Set-Join Conversion (SJC)
    SJC: Considering set-join conversion in SEL$2 (#0).
    Predicate Move-Around (PM)
    PM: Considering predicate move-around in SEL$1 (#0).
    PM:   Checking validity of predicate move-around in SEL$1 (#0).
    PM:   Passed validity checks.
    FPD: Considering simple filter push in SEL$1 (#0)
    FPD:   Current where clause predicates in SEL$1 (#0) :
             "VUNSCP"."DASFM"='30-apr-10'
    kkogcp: try to generate transitive predicate from check constraints for SEL$1 (#0)
    predicates with check contraints: "VUNSCP"."DASFM"='30-apr-10'
    after transitive predicate generation: "VUNSCP"."DASFM"='30-apr-10'
    finally: "VUNSCP"."DASFM"='30-apr-10'
    JPPD:     JPPD bypassed: View not on right-side of outer join
    FPD: Considering simple filter push in SEL$2 (#0)
    FPD:   Current where clause predicates in SEL$2 (#0) :
             apadrv-start: call(in-use=2936, alloc=16344), compile(in-use=38784, alloc=44568)
    kkoqbc-start
                : call(in-use=2936, alloc=16344), compile(in-use=40472, alloc=44568)
    kkoqbc-subheap (create addr=0x2a9740c1f0)
    Current SQL statement for this session:
    EXPLAIN PLAN FOR select * from VUNSCP where dasfm='30-apr-10'
    Peeked values of the binds in SQL statement
    PARAMETERS USED BY THE OPTIMIZER
      PARAMETERS WITH ALTERED VALUES
      _pga_max_size                       = 262140 KB
      cursor_sharing                      = similar
      _optimizer_cost_based_transformation = off
    Column Usage Monitoring is ON: tracking level = 1
    QUERY BLOCK TEXT
    EXPLAIN PLAN FOR select * from VUNSCP where dasfm='30-apr-10'
    QUERY BLOCK SIGNATURE
    qb name was generated
    signature (optimizer): qb_name=SEL$2 nbfros=1 flg=0
      fro(0): flg=0 objn=265023 hint_alias="X"@"SEL$2"
    SYSTEM STATISTICS INFORMATION
      Using NOWORKLOAD Stats
      CPUSPEED: 2503 millions instruction/sec
      IOTFRSPEED: 4096 bytes per millisecond (default is 4096)
      IOSEEKTIM: 10 milliseconds (default is 10)
    BASE STATISTICAL INFORMATION
    Table Stats::
      Table: SFMCPT  Alias:  X
        #Rows: 13036040  #Blks:  122880  AvgRowLen:  358.00
    Index Stats::
      Index: SFMCPT1  Col#: 2 3 4 8 5 7 118
        LVLS: 2  #LB: 58758  #DK: 13013072  LB/K: 1.00  DB/K: 1.00  CLUF: 11983641.00
      Index: SFMCPT2  Col#: 1
        LVLS: 2  #LB: 30031  #DK: 13483987  LB/K: 1.00  DB/K: 1.00  CLUF: 2410599.00
      Index: SFMCPT3  Col#: 3 4 8 5 7 2 118
        LVLS: 2  #LB: 39065  #DK: 13013072  LB/K: 1.00  DB/K: 1.00  CLUF: 12583891.00
    SINGLE TABLE ACCESS PATH
      BEGIN Single Table Cardinality Estimation
      Table: SFMCPT  Alias: X
        Card: Original: 13036040  Rounded: 13036040  Computed: 13036040.00  Non Adjusted: 13036040.00
      END   Single Table Cardinality Estimation
      Access Path: TableScan
        Cost:  27991.05  Resp: 27991.05  Degree: 0
          Cost_io: 26881.00  Cost_cpu: 33334822147
          Resp_io: 26881.00  Resp_cpu: 33334822147
      Best:: AccessPath: TableScan
             Cost: 27991.05  Degree: 1  Resp: 27991.05  Card: 13036040.00  Bytes: 0
    OPTIMIZER STATISTICS AND COMPUTATIONS
    GENERAL PLANS
    Considering cardinality-based initial join order.
    Permutations for Starting Table :0
    Join order[1]:  SFMCPT[X]#0
    WiF sort
        SORT resource      Sort statistics
          Sort width:         766 Area size:     1048576 Max Area size:   134215680
          Degree:               1
          Blocks to Sort:  108528 Row size:           68 Total Rows:       13036040
          Initial runs:         7 Merge passes:        1 IO Cost / pass:      58786
          Total IO sort cost: 167314      Total CPU sort cost: 16584848017
          Total Temp space used: 2093966000
    Best so far: Table#: 0  cost: 195857.3183  card: 13036040.0000  bytes: 677874080
    (newjo-stop-1) k:0, spcnt:0, perm:1, maxperm:80000
    Number of join permutations tried: 1
        SORT resource      Sort statistics
          Sort width:         766 Area size:     1048576 Max Area size:   134215680
          Degree:               1
          Blocks to Sort:  108528 Row size:           68 Total Rows:       13036040
          Initial runs:         7 Merge passes:        1 IO Cost / pass:      58786
          Total IO sort cost: 167314      Total CPU sort cost: 16584848017
          Total Temp space used: 2093966000
    Final - All Rows Plan:  Best join order: 1
      Cost: 195857.3183  Degree: 1  Card: 13036040.0000  Bytes: 677874080
      Resc: 195857.3183  Resc_io: 194195.0000  Resc_cpu: 49919670164
      Resp: 195857.3183  Resp_io: 194195.0000  Resc_cpu: 49919670164
    kkoipt: Query block SEL$2 (#0)
    ******* UNPARSED QUERY IS *******
    SELECT "X"."DASFM" "DASFM","X"."COINT" "COINT","X"."NUCPT" "NUCPT","X"."RGCOD" "RGCOD","X"."RGCID" "RGCID","X"."CODEV" "CODEV","X"."CTDEV" "CTDEV","X"."CDVRF" "CDVRF","X"."TXCHJ" "TXCHJ","X"."MTNLV" "MTNLV","X"."MTVDP" "MTVDP",DECODE(COUNT(*) OVER ( PARTITION BY "X"."COINT","X"."CODEV","X"."RGCOD","X"."CTDEV","X"."NUCPT","X"."CDVRF","X"."RGCID" ORDER BY "X"."DASFM" DESC  ROWS  BETWEEN 1 FOLLOWING  AND 1 FOLLOWING ),1,FIRST_VALUE("X"."MTNLV"+"X"."MTVDP") OVER ( PARTITION BY "X"."COINT","X"."CODEV","X"."RGCOD","X"."CTDEV","X"."NUCPT","X"."CDVRF","X"."RGCID" ORDER BY "X"."DASFM" DESC  ROWS  BETWEEN 1 FOLLOWING  AND 1 FOLLOWING ),NULL) "MTUNS",SUM("X"."MTNLV") OVER ( PARTITION BY "X"."COINT","X"."CODEV","X"."RGCOD") "MTNLI" FROM "SFMCPT" "X"
    kkoqbc-subheap (delete addr=0x2a9740c1f0, in-use=11856, alloc=12408)
    kkoqbc-end
              : call(in-use=57760, alloc=81816), compile(in-use=41096, alloc=44568)
    kkoqbc-start
                : call(in-use=57760, alloc=81816), compile(in-use=41184, alloc=44568)
    kkoqbc-subheap (create addr=0x2a9746b058)
    QUERY BLOCK TEXT
    select * from VUNSCP where dasfm='30-apr-10'
    QUERY BLOCK SIGNATURE
    qb name was generated
    signature (optimizer): qb_name=SEL$1 nbfros=1 flg=0
      fro(0): flg=1 objn=297481 hint_alias="VUNSCP"@"SEL$1"
    SYSTEM STATISTICS INFORMATION
      Using NOWORKLOAD Stats
      CPUSPEED: 2503 millions instruction/sec
      IOTFRSPEED: 4096 bytes per millisecond (default is 4096)
      IOSEEKTIM: 10 milliseconds (default is 10)
    BASE STATISTICAL INFORMATION
    Table Stats::
      Table: VUNSCP  Alias: VUNSCP  (NOT ANALYZED)
        #Rows: 0  #Blks:  0  AvgRowLen:  0.00
    OPTIMIZER STATISTICS AND COMPUTATIONS
    GENERAL PLANS
    Considering cardinality-based initial join order.
    Permutations for Starting Table :
    Join order[1]:  VUNSCP[VUNSCP]#0
    Best so far: Table#: 0  cost: 195857.3183  card: 13036040.0000  bytes: 1603432920
    (newjo-stop-1) k:0, spcnt:0, perm:1, maxperm:80000
    Number of join permutations tried: 1
    Final - All Rows Plan:  Best join order: 1
      Cost: 195857.3183  Degree: 1  Card: 13036040.0000  Bytes: 1603432920
      Resc: 195857.3183  Resc_io: 194195.0000  Resc_cpu: 49919670164
      Resp: 195857.3183  Resp_io: 194195.0000  Resc_cpu: 49919670164
    kkoipt: Query block SEL$1 (#0)
    ******* UNPARSED QUERY IS *******
    SELECT "VUNSCP"."DASFM" "DASFM","VUNSCP"."COINT" "COINT","VUNSCP"."NUCPT" "NUCPT","VUNSCP"."RGCOD" "RGCOD","VUNSCP"."RGCID" "RGCID","VUNSCP"."CODEV" "CODEV","VUNSCP"."CTDEV" "CTDEV","VUNSCP"."CDVRF" "CDVRF","VUNSCP"."TXCHJ" "TXCHJ","VUNSCP"."MTNLV" "MTNLV","VUNSCP"."MTVDP" "MTVDP","VUNSCP"."MTUNS" "MTUNS","VUNSCP"."MTNLI" "MTNLI" FROM  (SELECT "X"."DASFM" "DASFM","X"."COINT" "COINT","X"."NUCPT" "NUCPT","X"."RGCOD" "RGCOD","X"."RGCID" "RGCID","X"."CODEV" "CODEV","X"."CTDEV" "CTDEV","X"."CDVRF" "CDVRF","X"."TXCHJ" "TXCHJ","X"."MTNLV" "MTNLV","X"."MTVDP" "MTVDP",DECODE(COUNT(*) OVER ( PARTITION BY "X"."COINT","X"."CODEV","X"."RGCOD","X"."CTDEV","X"."NUCPT","X"."CDVRF","X"."RGCID" ORDER BY "X"."DASFM" DESC  ROWS  BETWEEN 1 FOLLOWING  AND 1 FOLLOWING ),1,FIRST_VALUE("X"."MTNLV"+"X"."MTVDP") OVER ( PARTITION BY "X"."COINT","X"."CODEV","X"."RGCOD","X"."CTDEV","X"."NUCPT","X"."CDVRF","X"."RGCID" ORDER BY "X"."DASFM" DESC  ROWS  BETWEEN 1 FOLLOWING  AND 1 FOLLOWING ),NULL) "MTUNS",SUM("X"."MTNLV") OVER ( PARTITION BY "X"."COINT","X"."CODEV","X"."RGCOD") "MTNLI" FROM "SFMCPT" "X") "VUNSCP" WHERE "VUNSCP"."DASFM"='30-apr-10'
    kkoqbc-subheap (delete addr=0x2a9746b058, in-use=11544, alloc=12408)
    kkoqbc-end
              : call(in-use=63208, alloc=81816), compile(in-use=41688, alloc=44568)
    apadrv-end: call(in-use=63208, alloc=81816), compile(in-use=42872, alloc=44568)
    sql_id=3xakq94fcx4td.
    Current SQL statement for this session:
    EXPLAIN PLAN FOR select * from VUNSCP where dasfm='30-apr-10'
    ============
    Plan Table
    ============
    ---------------------------------------+-----------------------------------+
    | Id  | Operation            | Name    | Rows  | Bytes | Cost  | Time      |
    ---------------------------------------+-----------------------------------+
    | 0   | SELECT STATEMENT     |         |       |       |  191K |           |
    | 1   |  VIEW                | VUNSCP  |   12M | 1529M |  191K |  00:39:11 |
    | 2   |   WINDOW SORT        |         |   12M |  646M |  191K |  00:39:11 |
    | 3   |    TABLE ACCESS FULL | SFMCPT  |   12M |  646M |   27K |  00:06:36 |
    ---------------------------------------+-----------------------------------+
    Predicate Information:
    1 - filter("DASFM"='30-apr-10')Edited by: Ahmed AANGOUR on 5 mai 2010 08:43

  • How to create variant for table/view ?

    Hi,
    When I go through SM30, I find a radio button called variant. I don't know the effect.
    Can anyone tell me how to create variant for table / view ?
    I want to know when we need to create variant for table/view.
    Best regards,
    Chris Gu

    hi ,
    Whenever you start a program in which selection screens are defined, the system displays a set of input fields for database-specific and program-specific selections. To select a certain set of data, you enter an appropriate range of values.
    For further information about selection screens, refer to Selection Screens in the ABAP User's Guide.
    If you often run the same program with the same set of selections (for example, to create a monthly statistical report), you can save the values in a selection set called a variant
    Procedure
    To create a new variant:
           1.      On the ABAP Editor initial screen, enter the name of the program for which you want to create a variant, select Variants, and choose Change.
           2.      On the variant maintenance initial screen, enter the name of the variant to be created.
    Note the naming convention for variants (see below).
           3.      Choose Create.
    If the program has more than one selection screen, a dialog box for screen assignment appears. The dialog box does not appear if the program only has one selection screen. The selection screen appears in this case.
           4.      If there is more than one selection screen, select the screens for which you want to create the variant
    5.      Choose Continue.
    The (first) selection screen for the report appears.
    If your program has more than one selection screen, use the scroll buttons in the left-hand corner of the application toolbar to navigate between them and to fill the fields with values. If you keep scrolling forwards, the Continue button appears on the last selection screen.
           6.      Enter the desired selection values, including multiple selection and dynamic selection.
           7.      Choose Continue.

  • How do I view data from OLAP DML program or statement?

    Hi
    I'm struggling to find any information on this topic anywhere on the net:
    How to view data produced by an Oracle OLAP DML program or statement.
    Background :
    I am using the latest AWM to create and maintain dimensions, cubes and measures. However, I've noticed that AWM, Discoverer OLAP, BI Beans, Excel Plug-In et al only provide a fixed set of formula and expressions.
    For example, I wanted to created a Standard Deviation (STDEV) on measure. After much searching I learnt that I could do this by creating a custom measure in AWM by importing a measure defined in an XML dcoument as it's not provided by AWM. The output in Discoverer OLAP on the surface seems to give the correct results, however when drilling from high level to lower level of time time dimension hierarchy Discoverer includes the parent value in the calculation and renders the values incorrect.
    My alternative is to use OLAP DML, as it provides a richer range of formuale and calculations possibilities and I was hoping to use this functionality to do create the measures. I have the distinct impression, however, that OLAP Worksheet is the only interface out there to do this.
    It seems very primitive and limited with regards to data output, almost like SQL*Plus in many ways.
    The burning questions I have are :
    1) Do I need to write DML programs to perform these custom measure calculations and add them to cubes or do I import the custom measures via AWM and XML?
    2) If the answer to question 1 is DML programming, is there a way of viewing the OLAP data output through a GUI interface without embedding the statements in SQL.
    3) Has anyone experienced or noticed the same problem with Discoverer OLAP, AWM Viewer that I mentioned above? i,e, statistical functions performed on whole dimension hierarchy rather than lower levels.
    I'm hoping someone out there knows what I am trying to do or get at. Keith, any ideas or suggestions?
    Thanks
    Kind Regards
    Greg

    Hi Greg,
    There are two ways to achieve what you want. Both require the use of custom calculations which cannot be created via the calculation wizard within AWM. To do this the easiest way is step outside of AWM and use the Excel Calculation utility that is on the OLAP OTN home page:
    Creating OLAP Calculations using Excel
    http://download.oracle.com/otn/java/olap/SpreadsheetCalcs_10203.zip
    Readme
    http://www.oracle.com/technology/products/bi/olap/OLAP_SpreadsheetCalcs.html
    (Oracle OLAP DML Reference contains a list of all the OLAP functions that can be used to create a custom calculation. http://www.oracle.com/technology/products/bi/olap/OLAP_DML_10.2.zip)
    The Excel tool lets you assign your own formula to a measure and then uploads that measure definition into your AW. It is possible to import a custom calculation template into AWM using the menu option on the Calculated Measure node of each cube, but it is safer to use the Excel Worksheet as this directly uses the public XML API and is always best to use a public API rather than hacking XML templates.
    There are two types of custom calculation:
    1) Formula: using a formula you can directly call any of the OLAP functions. For example to create a measure that returns the standard deviation for a measure then you would enter the following as the formula for your calculated measure:
    stddev(sales_revenue, time)
    The easiest way to test if the measure is working as expected is to use the data viewer within AWM.
    2) Programs - if the function you need is not provided or you want to do special processing to return a result you can create an OLAP DML program which can return a result. To do this use OLAP Worksheet to define the program and add a datatype to the DEFINE statement, such as :
    DEFINE PRG_SD_SALES PROGRAM DECIMAL
    In the program return the value you have calculated within your program. Within the program you can do just about anything but be aware that your program is going to fire for every cell within your virtual cube so it must be efficient and fast to execute. For a measure then you would enter the following as the formula for your calculated measure:
    prg_sd_sales(time, products, geographies, channels).
    The actual program code would look something like this:
    DEFINE PRG_SD_REVENUE PROGRAM DECIMAL
    PROGRAM
    argument T_TIME TIME
    argument T_PRODUCT PRODUCTS
    argument T_GEOGRAPHY GEOGRAPHIES
    argument T_CHANNEL CHANNELS
    variable D_RETURN decimal
    TEMPSTAT TIME, PRODUCTS, GEOGRAPHIES, CHANNELS
    do
    limit PRODUCTS to T_PRODUCT
    limit CHANNELS to T_CHANNEL
    limit GEOGRAPHIES to T_GEOGRAPHY
    limit TIME to T_TIME
    limit TIME add descendants using TIME_PARENTREL
    limit TIME keep TIME_LEVELREL 'CAL_MONTH'
    D_RETURN = stddev(SALES_REVENUE)
    doend
    return D_RETURN
    END
    This code computes a standard deviation the all months within the specified time range, at year level for all 12 months, at quarter level for the three months within the quarter ans returns NA at the month level.
    As the code executes within a implicit loop you can only change the status fo dimensions that for part of that implicit by using the TEMPSTAT command. To test the program at the command line use the OLAP Worksheet and the SHOW command to just return one cell of data, DO NOT SIMPLY USE THE RPR COMMAND because you will have to wait for the whole cube to be returned. If you want to use the RPR command make sure you limit your base dimensions first! You can use PRGTRACE to debug your program code it is not working as expected. Again, the easiest way to test if the measure is working as expected is to use the data viewer within AWM, although this will not return debug information so send all your debug output to a file to capture any errors and/or trace the execution flow.
    Hope this helps
    Keith Laker
    Oracle EMEA Consulting
    BI Blog: http://oraclebi.blogspot.com/
    DM Blog: http://oracledmt.blogspot.com/
    BI on Oracle: http://www.oracle.com/bi/
    BI on OTN: http://www.oracle.com/technology/products/bi/
    BI Samples: http://www.oracle.com/technology/products/bi/samples/

  • Statistical delivery date in Purchase Order

    Dear Friends.
    Please tell me, how system will display automatically statistical delivery date in PO while creation of PO for the first time? Because some times system will display this and some times it will not display even though i enter all the similar values. Hence i would like to know the conditions that will guide the system to calculate statistical delivery date automatically and display in PO. I checked and tried various options like entering/not entering/altering planned delivery time, GR processing time in material master purchasing view/MRP view and also in info record. Even with all these options i could not find the correct reason for automatic calculation of statistical delivery date in PO. I refered SAP  notes 1164767, 14422, 497150 but no use. Hence let me know the correct and right answer to my problem.

    Hello,
    Thanks for the answer. But as i already stated, Note 1164767 did not helped to solve the issue. I wanted to know how system will automaticall calculate and display statistical delivery date in PO? What are the parameters that govern this? What are the input values to calculate automatically statistical delivery date? Normally why this will be same as delivery date in PO? Suppose I want statistical delivery date = Planned delivery time + GR processing time + Purchasing department processing time + Some allowance (Say additional 1 day). For ex: Statistical Delivery date = 10 + 3 + 2 + 1= 16 days from the date of PO. Please tell me how to get this calculated in system automatically?
    Thanks in advance for your valuable answer.

  • Oracle 11G - Update is very slow on View

    I have big trouble with some Update query on Oracle 11G.
    I have a set of tables (5) of identical structures and a view that consists in an UNION ALL of the 5 tables.
    None of this table contains more than 20 000 rows.
    Let's call the view V_INTE_NE. Each of the basic table has a PRIMARY KEY defined on 3 NUMBERS(10,0) -> INTE_REF / NE_REF / INSTANCE.
    Now, I get 6 rows in another table and I want to update my view from the data of this small table (let's call it SMALL). This table has the 3 columns INTE_REF / NE_REF / INSTANCE.
    When I try to join the two tables :
    SELECT * FROM T_INTE_NE T2
    WHERE EXISTS ( SELECT 1 FROM SMALL T1 WHERE T2.INTE_REF = T1.INTEREF AND T2.NE_REF = T1.NEREF AND T2.INTE_INST = T1.INSTANCE )
    I get the 6 lines in 0.037 seconds
    When I try to update the view (I have an INSTEAD OF trigger that does nothing (just return for testing even without modifying anything), I execute the following query :
    UPDATE T_INTE_NE T2
    SET INTE_STATE = -11 WHERE
    EXISTS ( SELECT 1 FROM SMALL T1 WHERE T2.INTE_REF = T1.INTEREF AND T2.NE_REF = T1.NEREF AND T2.INTE_INST = T1.INSTANCE )
    The 6 rows are updated (at least TRIGGER is called) in 20 seconds.
    However, in the execution plan, I can't see where Oracle takes time to achieve the query :
    Plan hash value: 907176690
    | Id | Operation | Name | Rows | Bytes | Cost (%CPU)| Time |
    | 0 | UPDATE STATEMENT | | 6 | 36870 | 153 (1)| 00:00:02 |
    | 1 | UPDATE | T_INTE_NE | | | | |
    |* 2 | HASH JOIN RIGHT SEMI | | 6 | 36870 | 153 (1)| 00:00:02 |
    | 3 | TABLE ACCESS FULL | SMALL | 6 | 234 | 9 (0)| 00:00:01 |
    | 4 | VIEW | T_INTE_NE | 6 | 36636 | 143 (0)| 00:00:02 |
    | 5 | VIEW | X_V_T_INTE_NE | 6 | 18636 | 143 (0)| 00:00:02 |
    | 6 | UNION-ALL | | | | | |
    | 7 | TABLE ACCESS FULL| SECNODE1_T_INTE_NE | 1 | 3106 | 60 (0)| 00:00:01 |
    | 8 | TABLE ACCESS FULL| SECNODE2_T_INTE_NE | 1 | 3106 | 60 (0)| 00:00:01 |
    | 9 | TABLE ACCESS FULL| SECNODE3_T_INTE_NE | 1 | 3106 | 2 (0)| 00:00:01 |
    | 10 | TABLE ACCESS FULL| SECNODE4_T_INTE_NE | 1 | 3106 | 2 (0)| 00:00:01 |
    | 11 | TABLE ACCESS FULL| SECNODE5_T_INTE_NE | 1 | 3106 | 2 (0)| 00:00:01 |
    | 12 | TABLE ACCESS FULL| SYS_T_INTE_NE | 1 | 3106 | 17 (0)| 00:00:01 |
    Predicate Information (identified by operation id):
    2 - access("T2"."INTE_REF"="T1"."INTEREF" AND "T2"."NE_REF"="T1"."NEREF" AND
    "T2"."INTE_INST"="T1"."INSTANCE")
    Note
    - dynamic sampling used for this statement (level=2)
    Statistics
    3 user calls
    0 physical read total bytes
    0 physical write total bytes
    0 spare statistic 3
    0 commit cleanout failures: cannot pin
    0 TBS Extension: bytes extended
    0 total number of times SMON posted
    0 SMON posted for undo segment recovery
    0 SMON posted for dropping temp segment
    0 segment prealloc tasks
    What could explain the difference ?
    I get exactly the same execution plan (when autotrace is ON).
    Furthermore, if I try to do the same update on each of the basic tables, I get the rows updated instantaneously.
    Is there any reason for avoiding this kind of query ?
    Any help would be greatly appreciated :-)
    Regards,
    Patrick

    Sorry for this, I lost myself in conjonctures and I didn't think I would have to explain the whole case.
    So, I wrote a small piece of PL/SQL that reproduces the same issue.
    It seems that my issue is not due to the UPDATE but to the use of the IN predicate.
    As you can see at the end of the script, I try to join the 2 tables using different technics.
    The first query is very fast, the second is very slow.
    I need the second one if I want to do any update.
    DROP TABLE Part1;
    DROP TABLE Part2;
    DROP TABLE Part3;
    DROP TABLE Part4;
    CREATE TABLE Part1 ( Key1 NUMBER(10, 0), Key2 NUMBER(10, 0), Key3 NUMBER(10, 0), PartId NUMBER(10, 0) DEFAULT( 1 ) NOT NULL, Data1 VARCHAR2(1000), X_Data2 VARCHAR2(2000) NULL, X_Data3 VARCHAR2(2000) NULL, CONSTRAINT PK_Part1 PRIMARY KEY( Key1, Key2, Key3 ) );
    CREATE TABLE Part2 ( Key1 NUMBER(10, 0), Key2 NUMBER(10, 0), Key3 NUMBER(10, 0), PartId NUMBER(10, 0) DEFAULT( 2 ) NOT NULL, Data1 VARCHAR2(1000), X_Data2 VARCHAR2(2000) NULL, X_Data3 VARCHAR2(2000) NULL, CONSTRAINT PK_Part2 PRIMARY KEY( Key1, Key2, Key3 ) );
    CREATE TABLE Part3 ( Key1 NUMBER(10, 0), Key2 NUMBER(10, 0), Key3 NUMBER(10, 0), PartId NUMBER(10, 0) DEFAULT( 3 ) NOT NULL, Data1 VARCHAR2(1000), X_Data2 VARCHAR2(2000) NULL, X_Data3 VARCHAR2(2000) NULL, CONSTRAINT PK_Part3 PRIMARY KEY( Key1, Key2, Key3 ) );
    CREATE TABLE Part4 ( Key1 NUMBER(10, 0), Key2 NUMBER(10, 0), Key3 NUMBER(10, 0), PartId NUMBER(10, 0) DEFAULT( 4 ) NOT NULL, Data1 VARCHAR2(1000), X_Data2 VARCHAR2(2000) NULL, X_Data3 VARCHAR2(2000) NULL, CONSTRAINT PK_Part4 PRIMARY KEY( Key1, Key2, Key3 ) );
    CREATE OR REPLACE FUNCTION Decrypt
    x_in IN VARCHAR2
    ) RETURN VARCHAR2
    AS
    x_out VARCHAR2(2000);
    BEGIN
    SELECT REVERSE( x_in ) INTO x_out FROM DUAL;
    RETURN ( x_out );
    END;
    CREATE OR REPLACE VIEW AllParts AS
    SELECT Key1, Key2, Key3, PartId, Data1, Decrypt( X_Data2 ) AS Data2, Decrypt( X_Data3 ) AS Data3 FROM Part1
    UNION ALL
    SELECT Key1, Key2, Key3, PartId, Data1, Decrypt( X_Data2 ) AS Data2, Decrypt( X_Data3 ) AS Data3 FROM Part2
    UNION ALL
    SELECT Key1, Key2, Key3, PartId, Data1, Decrypt( X_Data2 ) AS Data2, Decrypt( X_Data3 ) AS Data3 FROM Part3
    UNION ALL
    SELECT Key1, Key2, Key3, PartId, Data1, Decrypt( X_Data2 ) AS Data2, Decrypt( X_Data3 ) AS Data3 FROM Part4;
    DROP TABLE Small;
    CREATE TABLE Small ( Key1 NUMBER(10, 0), Key2 NUMBER(10, 0), Key3 NUMBER(10, 0), Data1 VARCHAR2(1000) );
    BEGIN
    DECLARE
    n_Key NUMBER(10, 0 ) := 0;
    BEGIN
    WHILE ( n_Key < 50000 )
    LOOP
    INSERT INTO Part1( Key1, Key2, Key3 )
    VALUES( n_Key, FLOOR( n_Key / 10 ), FLOOR( n_Key / 100 ) );
    INSERT INTO Part2( Key1, Key2, Key3 )
    VALUES( n_Key, FLOOR( n_Key / 10 ), FLOOR( n_Key / 100 ) );
    INSERT INTO Part3( Key1, Key2, Key3 )
    VALUES( n_Key, FLOOR( n_Key / 10 ), FLOOR( n_Key / 100 ) );
    INSERT INTO Part4( Key1, Key2, Key3 )
    VALUES( n_Key, FLOOR( n_Key / 10 ), FLOOR( n_Key / 100 ) );
    n_Key := n_Key + 1;
    END LOOP;
    INSERT INTO Small( Key1, Key2, Key3, Data1 ) VALUES ( 1000, 100, 10, 'Test 1000' );
    INSERT INTO Small( Key1, Key2, Key3, Data1 ) VALUES ( 3000, 300, 30, 'Test 3000' );
    INSERT INTO Small( Key1, Key2, Key3, Data1 ) VALUES ( 5000, 500, 50, 'Test 5000' );
    COMMIT;
    END;
    END;
    SELECT T2.*
    FROM Small T1, AllParts T2
    WHERE T2.Key1 = T1.Key1 AND T2.Key2 = T1.Key2 AND T2.Key3 = T1.Key3;
    SELECT T1.*
    FROM AllParts T1
    WHERE ( T1.Key1, T1.Key2, T1.Key3 ) IN ( SELECT T2.Key1, T2.Key2, T2.Key3 FROM Small T2 );

  • Faq's in Interview point of view

    Hi All,
    I have a favour from all of you, I want to know what are the type of questions that would be asked in interview if I put these objects
    Reports :
    Created Material Master report for existing stock details.
    Created Material Master report for various storage locations.
    BDC :
    Migrated legacy data from Non SAP to SAP using SE11 transaction.
    Migrated legacy data using MK01 transaction.
    LSMW :
    Migrated legacy data using MM01 transaction.
    Reports
    Generated a report to print the purchase order details of goods received with purchase order number.
    Generated a program to create a report which calculates the material cost, it selects all the material issued for the entered service order number from stock.
    Sap Scripts
    Worked on Modifying the predefined Sap Script for Distributing Customer Billing.
    Smart Forms
    Worked on creating Job requisition form.
    Reports:
    Developed a ALV Grid Interactive report to evaluate the Vendor performance for a specified purchase organization in a particular period.
    Generated a ALV grid interactive report that displays the Material numbers and
    corresponding Maintenance status in the basic list and upon selection of a particular
    material no it gives the detailed information of that material no in two different levels.
    BDC:
    Interface to upload change the sales order for open sales order and partial sales order using the transaction VA02 through batch input method.
    SAP Scripts :
    Modified the Order layout set with slight changes in the terms and conditions, date
    format.
    Changes made to the invoice layout set as per the client requirement
    i would very much indebted to each and everyone whoever is helping to secure a job in ABAP..
    Thanking u in advance...
    srinivas
    u can also mail me @ [email protected]

    1.     How data is stored in cluster table?
    Each field of cluster table behaves as tables which contains the no. of entries.
    2. What are client dependant objects in abap/sap?
    SAP Script layout, text element, and some DDIC objects.
    3. On which even we can validate the input fields in module progams?
    In PAI (Write field statement on field you want to validate, if you want to validate group of fields put in chain and End chain statement.)
    4. In selection screen I have three fields, plant mat no and material group. If I input plant how do I get the mat no and material group based on plant dynamically?
    AT SELECTION-SCREEN ON VALUE-REQUEST FOR MATERIAL.
    CALL FUNCTION 'F4IF_INT_TABLE_VALUE_REQUEST' to get material and material group for the plant.
    5. How do you get output from IDOC?
    Data in IDOc is stored in segments, the output from Idoc is obtained by reading the data stored in its respective segments.
    6. When top of the page event is triggered?
    After excuteing first write statement in start-of-selection event.
    7. Can we create field without data element and how?
    In SE11 one option is available above the fields strip. Data element/ direct type.
    8. How do we debug sapscript?
    Go to SE71 give lay set name , go to utilities select debugger mode on.
    9. Which transaction code can I used to analyze the performance of ABAP program.
    TCode AL21.
    10. How can I copy a standard table to make my own z_table.
    Go to transaction SE11. Then there is one option to copy table. Press that button. Enter the name of the standard table and in the Target table enter Z table name and press enter.
    Following are some of the answers which I gave upto my knowledge.
    1. What is the use of 'outerjoin'
    Ans. With the use of outer join you can join the tables even there is no entry in all the tables used in the view.
    In case of inner join there should be an entry in al the tables use in the view.
    2. When to use logical database?
    Ans. Advantage of Logical databases:
    less coding s required to retrieve data compared to normal internel tables.
    Tables used LDB are in hierarchial structure.
    3. What is the use of 'table index'?
    Ans .Index is used for faster access of data base tables.
    4. What is the use of 'FOR ALL ENTRIES'?
    Ans. To avoid nested select statements we use SELECT FOR ALL ENTRIES statement.
    If there r more than 10000 records SELECT FOR ALL ENTRIES is used.
    Performance wise SELECT FOR ALL ENTRIES is better to use.
    5. Can you set up background processing using CALL TRANSACTION?
    Yes,Using No Screen Mode.
    6. What are table buffers?
    Table buffers reside locally on each application server in the system. The data of buffered tables can thus be accessed
    directly from the buffer of the application server. This avoids the time-consuming process of accessing the database.
    Buffering is useful if table needs to be accessed more no. of times in a program.
    ABAP Technical Interview Questions:
    1. What is the typical structure of an ABAP program?
    2. What are field symbols and field groups.? Have you used "component idx of structure" clause with field groups?
    3. What should be the approach for writing a BDC program?
    4. What is a batch input session?
    5. What is the alternative to batch input session?
    6. A situation: An ABAP program creates a batch input session. We need to submit the program and the batch session in background. How to do it?
    7. What is the difference between a pool table and a transparent table and how they are stored at the database level?
    8. What are the problems in processing batch input sessions? How is batch input process different from processing on line?
    9. What do you define in the domain and data element?
    10. What are the different types of data dictionary objects?
    11. How many types of tables exist and what are they in data dictionary?
    12. What is the step-by-step process to create a table in data dictionary?
    13. Can a transparent table exist in data dictionary but not in the database physically?
    14. What are the domains and data elements?
    15. Can you create a table with fields not referring to data elements?
    16. What is the advantage of structures? How do you use them in the ABAP programs?
    17. What does an extract statement do in the ABAP program?
    18. What is a collect statement? How is it different from append?
    19. What is open sql vs native sql?
    20. What does an EXEC SQL stmt do in ABAP? What is the disadvantage of using it?
    21. What is the meaning of ABAP editor integrated with ABAP data dictionary?
    22. What are the events in ABAP language?
    23. What is an interactive report? What is the obvious diff of such report compared with classical type reports?
    24. What is a drill down report?
    25. How do you write a function module in SAP? Describe.
    26. What are the exceptions in function module?
    27. What is a function group?
    28. How are the date abd time field values stored in SAP?
    29. What are the fields in a BDC_Tab Table?
    30. Name a few data dictionary objects?
    31. What happens when a table is activated in DD?
    32. What is a check table and what is a value table?
    33. What are match codes? Describe?
    34. What transactions do you use for data analysis?
    35. What is table maintenance generator?
    36. What are ranges? What are number ranges?
    37. What are select options and what is the diff from parameters?
    38. How do you validate the selection criteria of a report? And how do you display initial values in a selection screen?
    39. What are selection texts?
    40. What is CTS and what do you know about it?
    41. When a program is created and need to be transported to prodn does selection texts always go with it? if not how do you make sure? Can you change the CTS entries? How do you do it?
    42. What is the client concept in SAP? What is the meaning of client independent?
    43. Are programs client dependent?
    44. Name a few system global variables you can use in ABAP programs?
    45. What are internal tables? How do you get the number of lines in an internal table? How to use a specific number occurs statement?
    46. How do you take care of performance issues in your ABAP programs?
    47. What are datasets?
    48. How to find the return code of a stmt in ABAP programs?
    49. What are interface/conversion programs in SAP?
    50. Have you used SAP supplied programs to load master data?
    2. Adapted from response by Maram Roja on Tuesday, June 15, 2004
    1. What are the techniques involved in using SAP supplied programs? Do you prefer to write your own programs to load master data? Why?
    2. What are logical databases? What are the advantages/disadvantages of logical databases?
    3. What specific statements do you using when writing a drill down report?
    4. What are different tools to report data in SAP? What all have you used?
    5. What are the advantages and disadvantages of ABAP query tool?
    6. What are the functional areas? User groups? How does ABAP query work in relation to these?
    7. Is a logical database a requirement/must to write an ABAP query?
    8. What is the structure of a BDC sessions.
    9. What are Change header/detail tables? Have you used them?
    10. What do you do when the system crashes in the middle of a BDC batch session?
    11. What do you do with errors in BDC batch sessions?
    12. How do you set up background jobs in SAP? What are the steps? What are the event driven batch jobs?
    13. Is it possible to run host command from SAP environment? How do you run?
    14. What kind of financial periods exist in SAP? What is the relevant table for that?
    15. Does SAP handle multiple currencies? Multiple languages?
    16. What is a currency factoring technique?
    17. How do you document ABAP programs? Do you use program documentation menu option?
    18. What is SAPscript and layout set?
    19. What are the ABAP commands that link to a layout set?
    20. What is output determination?
    ABAP Interview Questions
    1.Without using Tcode SE11, How can we enter the values in to the table???
    2.What is the difference between Collect statement and Append Statement???
    3.What do you mean by correction and Transportation system???
    4.What is the difference between User Exits and BADI????
    5.How can we identify User exits in our screen???
    6.What do you mean by Inbound and Outbound interface???
    7.In realtime do we configure ALE systems or Administator will take care of that??
    8.How to release an object???
    9.What is the flow of a Sales document???
    10.What is the flow of Purchase order???
    12.What is the flow of Invoice???
    13.What are the standard IDOC's used???
    14.What do you mean by table control???Where will we use this???
    15.What are field symbols??Where will we use these symbols???
    Deepti
    1. There are other ways of entering data into a DB table. ex. B D C
    2. Collect statement collect/adds the records basing on a key field. allows to create summarised data sets.
    Append will append/add a record at the end of existing records
    8. to release an object - use se10/se9
    9. sales doc flow: S. A. - S. O. - Delivery - Billing
    12. Delivery - invoice.
    15. field symbols are used for dynamic allocation. at runtime u can assign a concrete field to field-symbol.
    Kishore
    1. you can go to abap editor (se38) and use insert statement for insertion update for update and modify for modifications.
    2. collect will not allow duplicate entries, while append can allow duplicates.
    3. if any changes are made to objects they are to be transported to different systems i.e, change and transport.
    4. in user exits we go by general method for enhancements while BADIs we use objects (oops concepts)
    methods for enhancement.
    14. table controls are enhanced version for step loops where we can expand rows .main difference between these two
    is step loop can be expanded to two lines table controls can`t.
    15. field symbols are just like pointers concept which are used in C language. We use them when we want to refer to
    the fields considered,it doesnt allocate any memory for it.
    Venu Rapolu
    1. Ans: Using BDC., LSMW,,ALE., BAPI
    2. Ans: Collect: it adds the numeric fields to the existing non numeric key field records., thereby avoiding duplicate
    values., and append will simply adds the record
    3 . Ans: The CTS contains work bench organizer and transport system :
    The workbench organizer is used to record and contol changes to the ABAP/4 development objects;
    The transport system is used to move objects from an SAP dev.system to production system
    6. Ans: Outbound interface is used to send IDocs to the ALE server.,
    Inbound interface is used to Analyse the received Idoc.
    7. Ans: We (ABAPers) don't do.
    8. Ans: se09 or se10
    13. Ans: MATMAS
    CREMAS
    DEBMAS
    GLMAST etc...
    14. Ans: to display records in table format., we use in Screens
    15. Ans: we assign the field content at run time with ASSIGN stmt.
    Satish D
    1. goto se16 u can view the table contents
    2. collect will collects all the numeric fields of nonnumeric field values
    append will append record at the end of table
    3. when we are creatiing new task like dev. a new prg.. after completion of dev. that will be trnsported to testing system or production system to do that we are assigning an
    transport request from our dev.class(system) by using se09
    4. badis are dev. by class and inheritence methods where as user exitsdev by SAP only and empty shells filled wit user logic
    5. go to system then status
    6. outbound generating an idoc in ale layer with master_idoc_create_messgetype,
    inbound is receivng system with idoc_inbound_process, gives an return variable wether it is sucess or not
    7. no we have to configur that
    8. go to se09 or se01 there write your task no and use release button it will asks whats the other system name
    and number enter them and relase by pressing jeep button
    9. inquiry
    quotation
    sales order
    shipping
    delivery
    invoice
    return goods
    12. after creation of delivry note invoice is prepared
    13. master_idoc_distribute will generate standard idoc
    Deepti
    I am enclosing some of answers which I know.
    1. We can use t.Code SE16 to enter values into table only if table maintainence is allowed for that table.
    2. Append will add new entries into the table where as collect add into the numeric type fields if other charatcer fields
    matches to your selection criteria.
    3. CTS used for creation of ABAP development transport requests.The transaction code for this is SE10.
    4. UserExits r used for adding additional functionality to the existing SAP standard transactions.Using UserExits we can add additional functionality standard SAP functionality without making any changes to the original code.BADI is a standardized interface for ABAP sources that enables partners and customers to enhance SAP-delivered programs in their namespace.
    5 .We can identify User exits by using transactions CMOD and SMOD.
    6. After entering transaction code SE10 select the transport request which u want to transport and click on transport icon(Truck symbol) to release it.
    10. Purchase Requistion->RFQ->Vendor Evaluation->Purchase Order(ME21).
    14. Table control is the only facility provide thru dialog programming when we come acrosse the use of updating standared,deletion,insertion and all database operations.
    15. Field symbols r pointers to the existing data types(like 'C') which does not allocate any memory space. These are used faster access of data.
    Answers to some ABAP Interview Questions:
    Questions which I have faced in an interview:
    1) What is runtime analysis? Have you used this?
    2) What is meant by performance analysis? Have done anything to improve the performance?
    3) How to transfer the objects? Have to transferred any objects?
    4) How did you test the developed objects?
    5) What is the difference between SAP Memory and ABAP Memory?
    6) In order to upload Purchase order details, how you handle multiple values for a single field?
    Eg: Item field may contain no. of values for a record
    7) What is the procedure you followed to upload the data?
    8) How did you handle errors in Call Transaction?
    9) Among the Call Transaction and Session Method, which is faster?
    10) What are the difference between Interactive and Drill Down Reports?
    11) How to pass the variables to forms?
    12) How to create a link between modified form and modified print program?
    13) What is the table, which contain the details of all the name of the programs and forms?
    14) How did you test the form u developed? How did you taken print?
    15) What are Standard Texts?
    16) What is the difference between Clustered Tables and Pooled Tables?
    17) What is pf-status?
    18) Among "Move" and "Move Corresponding", which is efficient one?
    19) What are the output type and Tcodes?
    20) Where we use Chain and Endchain?
    21) Do you use select statement in loop endloop, how will be the performance? To improve the performance?
    22) In select-options, how to get the default values as current month first date and last date by default?
    Eg: 1/12/2004 and 31/12/2004
    Go thru these answers:
    1) What is runtime analysis? Have you used this?
    It's checks program execution time in microseconds. When you go to se30.if you give desired program name in performance file. It will take you to below screen. You can get how much past is your program.
    2) What is meant by performance analysis? Have done
    3) How to transfer the objects? Have you transferred any objects?
    4) How did you test the developed objects?
    I was testing a developed object. There are two types of testing
    - Negative testing
    - Positive testing
    In negative testing we will give negative data in input and we check any errors occurs.
    In positive testing we will give positive data in input for checking errors.
    8) How did you handle errors in Call Transaction?
    We can create a internal table like 'bsgmcgcoll'. All the messages will go to internal table. We can get errors in this internal table.
    Below messages are go to internal table. when you run the call transaction.
    - Message type
    - Message id
    - Message Number
    - Variable1
    - Variable2
    - Variable3
    9) Among the Call Transaction and Session Method, which is faster?
    Call transaction is faster then session method. But usually we use session method in real time...because we can transfer large amount of data from internal table to database and if any errors in a session. Process will not complete until session get correct.
    10) What are the difference between Interactive and Drill Down Reports?
    ABAP/4 provides some interactive events on lists such as AT LINE-SELECTION (double click) or AT USER-COMMAND (pressing a button). You can use these events to move through layers of information about individual items in a list.
    Drill down report is nothing but interactive report...drilldown means above paragraph only.
    11) How to pass the variables to forms?
    12) What is the table, which contain the details of all the name of the programs and forms?
    Table contains vertical and horizontal lines. We can store the data in table as blocks. We can scroll depends upon your wish. And these all are stored in database (data dictionary).
    Which contain the details of all the name of the programs and forms? (I don't know).
    13) How did you test the form u developed? How did you taken print?
    14) What are Standard Texts?
    16) What is the difference between Clustered Tables and Pooled Tables?
    A pooled table is used to combine several logical tables in the ABAP/4 dictionary. Pooled tables are logical tables that must be assigned to a table pool when they are defined.
    Cluster table are logical tables that must be assigned to a table cluster when they are defined.
    Cluster table can be used to store control data they can also used to store temporary data or text such as documentation.
    17) What is pf-status?
    Pf status is used in interactive report for enhancing the functionality. If we go to se41, we can get menus, items and different function keys, which we are using for secondary list in interactive report.
    18) Among "Move" and "Move Corresponding", which is efficient one?
    I guess, 'move corresponding' is very efficient then 'move' statement. Because usually we use this stamtent for internal table fields only...so if we give move corresponding. Those fields only moving to other place (what ever you want).
    19) What are the output type and Tcodes?
    20) Where we use Chain and End chain?
    21) Do you use select statement in loop end loop, how will be the performance? To improve the performance?
    22) In select-options, how to get the default values as current month first date and last date by default?
    Eg: 1/12/2004 and 31/12/2004
    SAP ABAP interview questions
    Thanks to the reader who sent in this question set:
    1. What is an ABAP data dictionary?- ABAP 4 data dictionary describes the logical structures of the objects used in application development and shows how they are mapped to the underlying relational database in tables/views.
    2. What are domains and data element?- Domains:Domain is the central object for describing the technical characteristics of an attribute of an business objects. It describes the value range of the field. Data Element: It is used to describe the semantic definition of the table fields like description the field. Data element describes how a field can be displayed to end-user.
    3. What is foreign key relationship?- A relationship which can be defined between tables and must be explicitly defined at field level. Foreign keys are used to ensure the consistency of data. Data entered should be checked against existing data to ensure that there are now contradiction. While defining foreign key relationship cardinality has to be specified. Cardinality mentions how many dependent records or how referenced records are possible.
    4. Describe data classes.- Master data: It is the data which is seldomly changed. Transaction data: It is the data which is often changed. Organization data: It is a customizing data which is entered in the system when the system is configured and is then rarely changed. System data:It is the data which R/3 system needs for itself.
    5. What are indexes?- Indexes are described as a copy of a database table reduced to specific fields. This data exists in sorted form. This sorting form ease fast access to the field of the tables. In order that other fields are also read, a pointer to the associated record of the actual table are included in the index. Yhe indexes are activated along with the table and are created automatically with it in the database.
    6. Difference between transparent tables and pooled tables.- Transparent tables: Transparent tables in the dictionary has a one-to-one relation with the table in database. Its structure corresponds to single database field. Table in the database has the same name as in the dictionary. Transparent table holds application data. Pooled tables. Pooled tables in the dictionary has a many-to-one relation with the table in database. Table in the database has the different name as in the dictionary. Pooled table are stored in table pool at the database level.
    7. What is an ABAP/4 Query?- ABAP/4 Query is a powerful tool to generate simple reports without any coding. ABAP/4 Query can generate the following 3 simple reports: Basic List: It is the simple reports. Statistics: Reports with statistical functions like Average, Percentages. Ranked Lists: For analytical reports. - For creating a ABAP/4 Query, programmer has to create user group and a functional group. Functional group can be created using with or without logical database table. Finally, assign user group to functional group. Finally, create a query on the functional group generated.
    8. What is BDC programming?- Transferring of large/external/legacy data into SAP system using Batch Input programming. Batch input is a automatic procedure referred to as BDC(Batch Data Communications).The central component of the transfer is a queue file which receives the data vie a batch input programs and groups associated data into “sessions”.
    9. What are the functional modules used in sequence in BDC?- These are the 3 functional modules which are used in a sequence to perform a data transfer successfully using BDC programming: BDC_OPEN_GROUP - Parameters like Name of the client, sessions and user name are specified in this functional modules. BDC_INSERT - It is used to insert the data for one transaction into a session. BDC_CLOSE_GROUP - This is used to close the batch input session.
    10. What are internal tables?- Internal tables are a standard data type object which exists only during the runtime of the program. They are used to perform table calculations on subsets of database tables and for re-organising the contents of database tables according to users need.
    11. What is ITS? What are the merits of ITS?- ITS is a Internet Transaction Server. ITS forms an interface between HTTP server and R/3 system, which converts screen provided data by the R/3 system into HTML documents and vice-versa. Merits of ITS: A complete web transaction can be developed and tested in R/3 system. All transaction components, including those used by the ITS outside the R/3 system at runtime, can be stored in the R/3 system. The advantage of automatic language processing in the R/3 system can be utilized to language-dependent HTML documents at runtime.
    12. What is DynPro?- DynPro is a Dynamic Programming which is a combination of screen and the associated flow logic Screen is also called as DynPro.
    13. What are screen painter and menu painter?- Screen painter: Screen painter is a tool to design and maintain screen and its elements. It allows user to create GUI screens for the transactions. Attributes, layout, filed attributes and flow logic are the elements of Screen painter. Menu painter: Menu painter is a tool to design the interface components. Status, menu bars, menu lists, F-key settings, functions and titles are the components of Menu painters. Screen painter and menu painter both are the graphical interface of an ABAP/4 applications.
    14. What are the components of SAP scripts?- SAP scripts is a word processing tool of SAP which has the following components: Standard text. It is like a standard normal documents. Layout sets. - Layout set consists of the following components: Windows and pages, Paragraph formats, Character formats. Creating forms in the R/3 system. Every layout set consists of Header, paragraph, and character string. ABAP/4 program.
    15. What is ALV programming in ABAP? When is this grid used in ABAP?- ALV is Application List viewer. Sap provides a set of ALV (ABAP LIST VIEWER) function modules which can be put into use to embellish the output of a report. This set of ALV functions is used to enhance the readability and functionality of any report output. Cases arise in sap when the output of a report contains columns extending more than 255 characters in length. In such cases, this set of ALV functions can help choose selected columns and arrange the different columns from a report output and also save different variants for report display. This is a very efficient tool for dynamically sorting and arranging the columns from a report output. The report output can contain up to 90 columns in the display with the wide array of display options.
    16. What are the events in ABAP/4 language?- Initialization, At selection-screen, Start-of-selection, end-of-selection, top-of-page, end-of-page, At line-selection, At user-command, At PF, Get, At New, At LAST, AT END, AT FIRST.
    17. What is CTS and what do you know about it?- The Change and Transport System (CTS) is a tool that helps you to organize development projects in the ABAP Workbench and in Customizing, and then transport the changes between the SAP Systems and clients in your system landscape. This documentation provides you with an overview of how to manage changes with the CTS and essential information on setting up your system and client landscape and deciding on a transport strategy. Read and follow this documentation when planning your development project.
    18. What are logical databases? What are the advantages/ dis-advantages of logical databases?- To read data from a database tables we use logical database. A logical database provides read-only access to a group of related tables to an ABAP/4 program. Advantages: i)check functions which check that user input is complete, correct,and plausible. ii)Meaningful data selection. iii)central authorization checks for database accesses. iv)good read access performance while retaining the hierarchical data view determined by the application logic. dis advantages: i)If you donot specify a logical database in the program attributes,the GET events never occur. ii)There is no ENDGET command,so the code block associated with an event ends with the next event statement (such as another GET or an END-OF-SELECTION).
    19. What is a batch input session?- BATCH INPUT SESSION is an intermediate step between internal table and database table. Data along with the action is stored in session ie data for screen fields, to which screen it is passed, program name behind it, and how next screen is processed.
    20. How to upload data using CATT ?- These are the steps to be followed to Upload data through CATT: Creation of the CATT test case & recording the sample data input. Download of the source file template. Modification of the source file. Upload of the data from the source file.
    21. What is Smart Forms?- Smart Forms allows you to create forms using a graphical design tool with robust functionality, color, and more. Additionally, all new forms developed at SAP will be created with the new Smart Form solution.
    22. How can I make a differentiation between dependent and independent data?- Client dependent or independent transfer requirements include client specific or cross client objects in the change requests. Workbench objects like SAPscripts are client specific, some entries in customizing are client independent. If you display the object list for one change request, and then for each object the object attributes, you will find the flag client specific. If one object in the task list has this flag on, then that transport will be client dependent.
    23. What is the difference between macro and subroutine? - Macros can only be used in the program the are defined in and only after the definition are expanded at compilation / generation. Subroutines (FORM) can be called from both the program the are defined in and other programs . A MACRO is more or less an abbreviation for some lines of code that are used more than once or twice. A FORM is a local subroutine (which can be called external). A FUNCTION is (more or less) a subroutine that is called external. Since debugging a MACRO is not really possible, prevent the use of them (I’ve never used them, but seen them in action). If the subroutine is used only local (called internal) use a FORM. If the subroutine is called external (used by more than one program) use a FUNCTION.
    Please check these links.
    http://www.sap-img.com/abap/abap-interview-question.htm
    http://www.sap-img.com/abap/answers-to-some-abap-interview-questions.htm
    http://sap.ittoolbox.com/documents/document.asp?i=3240
    http://www.techinterviews.com/?p=198
         http://www.sapbrain.com/FAQs/TECHNICAL/SAP_ABAP_DATADICTIONARY_FAQ.html
    http://www.****************/InterviewQ/interviewQ.htm
    http://help.sap.com/saphelp_46c/helpdata/en/35/2cd77bd7705394e10000009b387c12/frameset.htm
    http://www.techinterviews.com/?p=198
    http://www.techinterviews.com/?p=326
    http://www.sap-img.com/abap/answers-to-some-abap-interview-questions.htm
    http://www.sap-img.com/abap/more-than-100-abap-interview-faqs.htm
    http://www.geekinterview.com/Interview-Questions/SAP-R-3/ABAP
    http://sap.ittoolbox.com/documents/popular-q-and-a/abap-sample-interview-questions-3240
    http://www.sap-img.com/abap/abap-interview-question.htm
    http://www.allinterview.com/Interview-Questions/ABAP.html
    regards,
    srinivas

Maybe you are looking for