Retrive SQL from Webi report and process it for large volume of data

We have a Scenario where, we need to extract large volumes of data into flat files and distribute them from u2018Teradatau2019 warehouse which we usually call them as u2018Extractsu2019. But the requirement is such that, Business users wants to build their own u2018Adhoc Extractsu2019.   The only way, I thought, to achieve this, is Build a universe, create the query, save the report and do not run it. Then write a RAS SDK to retrieve the SQL code from the reports and save into .txt file and process it directly in Teradata?
Is there any predefined Solution available with SAP BO or any other tool for this kind of Scenarios?

Hi Shawn,
Do we have some VB macro to retrieve Sql queries of data providers of all the WebI reports in CMS.
Any information or even direction where I can get information will be helpful.
Thanks in advance.
Ashesh

Similar Messages

  • Transport Webi Reports and Web services as the backend of dashboards from BO Production to BO Development system

    Hello Experts,
    I am working on SAP BO 4.1. I have made several dashboards on top of web services ie;Web Service Method. I have 2 systems in BO ie; Development and Production Systems.The BW production system is connected to BO Development and Production both.
    The Webi reports are made on top of BI BEx Query. From the webi reports, BI Web Services are made on top of which the dashboards are made further.
    The Webi Reports, Web Services and the Dashboards everything is made directly in BO Production.
    My question is, Can I transport the Webi Reports and the Web Services from BO Production to BO Development?
    And If yes, will it have any other impact on webi reports, web services or dashboards?
    Thanks & Regards,
    Alfred Thomas

    Hi Gill,
    As per your reply,with the promotion managament i have make the web services again manually in Dev system...Right?
    Is there any way possible that i can transport the webservices and the webi reports usind Query AS A Web Service Designer. And if yes, through this QAAWS will the WSDL or the URL required for the web services in the connection button in dashboards will automatically updated or changed as per the Development System?
    But i am not able to enable the "Deploy to Other servers Option" in QAAWS.
    Can you please help?
    Regards,
    Alfred thomas

  • Scheduled a webi report and set it send to EMAIL.

    Our customer found when they scheduled a webi report and set it send to EMAIL, it will cost much more time than usual after they have their SIA restarts.
    The first time it takes 4-5 times longer to receive a report in mailbox than next time onwards.
    Please help.
    Thanks Much,
    Abhishek

    Hi Abishek,
    If the schedule is taking 4-5 times longer to schedule following the SIA restart I suppose it is possible that the Webi Job Server and Webi Processing Servers are not fully registered with the CMS - can you check that.
    Also can you test by ensuring that if there are any Webi Job Servers\Webi Processing Servers in the environment which are stopped\disabled, can you retest after deleting those servers. There is a bug with XI3.1:
    ADAPT01371367
    Description:
    Scheduled jobs often have a delay when the system is upgraded to BusinessObjects Enterprise XI 3.1 SP2 FP 2.2 or FP 2.3.This problem also happens when some of the servers required for scheduling are stopped and disabled. In this case, the Adaptive Job Server (AJS) tries to confirm if all available Webi Report Servers are running, and this causes the delay in running the scheduled jobs.
    New Behavior:
    This problem has been resolved.
    Do you have a test system where you could install FP3.1 and check the fix?
    The only "workaround" I could suggest would be to ensure that all services needed for the schedule are running\enabled (Adaptive Job Servers and Webi Processing Servers). This will stop the Adaptive Job Server from its continual querying for those servers. The adverse effect of this is that the more unnecessary services which are running will necessitate increased network communication.
    To check if this is the actual issue you are hitting, you would need to trace the schedule and check for messages like this in the logs:
    [2010-08-27 12:02:08,769] [TID:24] [ERROR] [webi.PublishingHelper.<init>():89]: unable to retrieve all mandatories servers to proceed this job
    Alternatively, log a message with SAP Support for further investigation.
    <<Text removed>>
    Kind regards,
    John
    Edited by: Matt on Oct 10, 2011 10:14 AM

  • Export graphs from Web report

    Hi.
    I need to copy graphs from web reports in 3.5.To do this, i have followed a roundabout:
    1. Save the page as mht (web archive file)
    2.Copy the image from there and paste into Powerpoint or wherever required.
    Is there any better way to do it? Or can I bring the save as mht option in the context menu?
    Is SAP addressing it any soon?
    Also is there any way to send graphs in addition to the tables via the 'distribute as email option"?
    Cheers
    Anand

    When you broadcast using the standard template the template '0QUERY_TEMPLATE_BROADCASTING' is used for broadcasting. This template only has a navigation block and table item, this is the reason why the chart is not displayed in this template. If you create a custom template for broadcasting which has chart item in it you will be able to get the chart displayed (it will be always displayed.) See the attached link for custmoizing.
    http://help.sap.com/saphelp_nw04/helpdata/en/3a/0e044017355c0ce10000000a1550b0/content.htm
    You can also override this behaviour if you attach the query to a custom web template, in this case when you broadcast you will get exactly what you see in the current report.

  • How to develp a subreport in the same page of the main report and processed when is called

    SSRS 2012
    Hi guys,
    I am developing a report that includes a subreport.
    therefore The report is made up by 2 sections:
    1) Show a grid (several rows - may be 100 or more - and some columns). It represent high level info  (Each row represent a specific item)
    2) The user click on a Row/column (eg: Item1, to see details that must be shown as graph.
    Here the problem: Searching info about SSRS reporting for subreports, drilldown, nested report and so on, no ones accomplish the requirements for the following reasons:
    1) Subreports: NOT open in the same page. The user must <go back> to see detail for other items, such as Item 2....
    2) Nested and DrillDown: Process data as the Main report (they are hidden and displayed after user action). In my case this leads to performance problem, because: if the main report has 200 objects (it means 200 graphs processed and hidden at the same time
    for the main reports).
    3) DrillThrough and Subreport: The detail data are processed only when the user select the interested item in the main report. Here the problem: the report is shown externally.
    Please, it there any possibility to have a subreport (or anything else) to be executed in the same report as the main report and processed only when the user make action?
    Thanks for your help.

    Hi Fasttrack2,
    According to your description, there is a main report with more than 100 rows of summary information, when users click the item of the row, the report need to jump to detail information shows as graph. The problem you are facing is that you are not sure
    subreport, drilldown or drillthrough report should be used? You hope to set display detail information in the main report and keep high performance of report.
    In Reporting Services, each subreport instance is a separate query execution and a separate report processing task. Subreports are recommended when there are just a few subreport instances. We should not use subreports inside a group when there are many
    group instances, instead, consider using drillthrough reports. Drilldown reports process all data even when the data is first hidden. In order to improve the performance of the report, we can split drilldown reports into parameterized drillthrough reports,
    because Drillthrough reports do not run until a user clicks the drillthrough link in the main report.
    In this case, we can consider using drilldown or drillthrough report. In drilldown report, when we click a plus or minus button to expand or collapse a section of a report, the detail data will be displayed in place. By using drillthrough report, when we
    click the link in the main report, the detail data will be displayed, we can come back to the main report by click return button in the drillthrough report.
    Reference:
    Troubleshooting Reports: Report Performance
    Drillthrough, Drilldown, Subreports, and Nested Data Regions
    If you have any more questions, please feel free to ask.
    Best Regards,
    Wendy Fu
    If you have any feedback on our support, please click
    here.

  • Call tcode from alv report and passing  group of values

    hi all .
    i want to call tcode from alv report and passing an internal table or group of values to a selection option of that t code ? how
    ex. passing group of GL to fbl3n and display the detials of all .
    thank you

    Dear,
    You have done a small mistake
    --> rspar_line-option = 'EQ'.
         rspar_line-HIGH = PDATE-HIGH.
    u r passing "high" value and in "option u r passing "EQ" so how it will work!!!
    So if u r passing only 1 date or more dates like 01.01.2010 , 15.02.2010 , 10.03.2010 then pass
    rspar_line-selname = 'SO_BUDAT'.
    rspar_line-kind = 'S'.
    rspar_line-sign = 'I'.
    rspar_line-option = 'EQ'.
    rspar_line-LOW = PDATE-HIGH.
    APPEND rspar_line TO rspar_tab.
    or if u r passing low & high date means in range like 01.01.2010 to 30.01.2010, then pass
    rspar_line-selname = 'SO_BUDAT'.
    rspar_line-kind = 'S'.
    rspar_line-sign = 'I'.
    rspar_line-option = 'BT''.
    rspar_line-LOW = PDATE-LOW.
    rspar_line-HIGH = PDATE-HIGH.
    APPEND rspar_line TO rspar_tab.
    try above code , hope it helps...
    i think u cannot use "call transaction using bdcdata" in ur case bcoz as u said in ur 1st post u want to display the details of all but still if u want to use then u should pass all parameters in  loop.
    PROGRAM
    DYNPRO
    DYNBEGIN
    FNAM
    FVAL
    ex:-
    LOOP AT GT_TEMP INTO GS_TEMP.
    CLEAR bdcdata_wa.
    bdcdata_PROGRAM = 'SAPXXXX'.
    bdcdata_DYNPRO = '1000'.
    bdcdata_DYNBEGIN = 'X'.
    bdcdata_wa-fnam = '''.
    bdcdata_wa-fval = ''.
    APPEND bdcdata_wa TO bdcdata_tab.
    CLEAR bdcdata_wa.
    bdcdata_PROGRAM = ''.
    bdcdata_DYNPRO = ''.
    bdcdata_DYNBEGIN = ''.
    bdcdata_wa-fnam = 'SD_SAKNR'.
    bdcdata_wa-fval = GS_TEMP-GLACCOUNT.
    APPEND bdcdata_wa TO bdcdata_tab.
    CLEAR bdcdata_wa.
    bdcdata_PROGRAM = ''.
    bdcdata_DYNPRO = ''.
    bdcdata_DYNBEGIN = ''.
    bdcdata_wa-fnam = 'BDC_OKCODE'.
    bdcdata_wa-fval = 'XXX'.
    APPEND bdcdata_wa TO bdcdata_tab.
    ENDLOOP.
    try above code if u r using call transaction...
    Edited by: mihir6666 on Jul 9, 2011 3:10 PM
    Edited by: mihir6666 on Jul 9, 2011 3:11 PM
    Edited by: mihir6666 on Jul 9, 2011 3:13 PM

  • To Link a Crystal Report (2008) from Webi report - Prompts not to Repeat

    Hi ,
    I have created a Webi report and Crystal report (2008) both have the same prompts.now i want to create a Hyperlink in Webi to link the Crystal report.
    i am able to do this. however the crystal report is again asking to provide the prompts which i already given in Webi.which should not happen.
    So how to achieve this scenario in Webi.        
    Need your help as soon as possible.
    Regards,
    Kishore.

    Hi
    the information you need are in the following manuals
    [Viewing Crystal Reports Using URL Reporting|http://help.sap.com/businessobject/product_guides/boexir4/en/xi4_cr_url_reporting_en.pdf]
    and
    [Viewing Documents Using OpenDocument|http://help.sap.com/businessobject/product_guides/boexir4/en/xi4_opendocument_en.pdf]
    Thanks!

  • Passing parameter from WEBI report to xcelsius

    hi
    i transfered the parameter from WEBI report (3.1 sp2) table using a link to a gauhe from xcelsius
    i can see the parameter the right number in the gauge value but the problem is that the pin does not move

    How are you connecting your report to Xcelsius? Live Office or QAAWS?

  • Unable to create a webi report and 3 times it prompting for logon credentai

    Hi,
    We have a issue,
    1.We are  unable to create a webi report and 3 times it is prompting for logon credentials.
    3.After giving the logon credentials 3 times then it is working fine.
    4.We are  using IIS Jakarta redirector .
    5.We are  able to view the webi documents successfully ,the issue occurs only when we  create a new webi report.
    Any one please help me on this issue.
    Thank you in advance.
    Environment u2013
    BOXI3.1,
    SSO,
    IIS Jakarta redirector.
    Thanks & Regards,
    Powell.

    bypass IIS and jakarta (go to tomcat directly with an enterprise account like administrator), does the issue still exist?
    Regards,
    Tim

  • How to download data in excel from web report in sap ?

    how to download data in excel from web report made  in sap abap?
    through tcode smw0.

    for exemple using
    MS_EXCEL_OLE_STANDARD_DAT

  • Re : Web reporting and WAD

    Hello friends,
    I need to give a presentation on Web reporting and WAD. Can anybody share the relavant documents or ppts.
    I dont have much time to prepare.
    Thanks in advance.
    Thanks
    Bujji

    HI,
    Go through below link to get idea on new WAD features
    http://help.sap.com/saphelp_nw2004s/helpdata/en/29/4d15422ecce02ce10000000a1550b0/frameset.htm
    http://help.sap.com/saphelp_nw70/helpdata/EN/00/e8d13f7fb44c21e10000000a1550b0/frameset.htm
    Regards,
    rik

  • Processing large volumes of data in PL/SQL

    I'm working on a project which requires us to process large volumes of data on a weekly/monthly/quarterly basis, and I'm not sure we are doing it right, so any tips would be greatly appreciated.
    Requirement
    Source data is in a flat file in "short-fat" format i.e. each data record (a "case") has a key and up to 2000 variable values.
    A typical weekly file would have maybe 10,000 such cases i.e. around 20 million variable values.
    But we don't know which variables are used each week until we get the file, or where they are in the file records (this is determined via a set of meta-data definitions that the user selects at runtime). This makes identifying and validating each variable value a little more interesting.
    Target is a "long-thin" table i.e. one record for each variable value (with numeric IDs as FKs to identify the parent variable and case.
    We only want to load variable values for cases which are entirely valid. This may be a merge i.e. variable values may already exist in the target table.
    There are various rules for validating the data against pre-existing data etc. These rules are specific to each variable, and have to be applied before we put the data in the target table. The users want to see the validation results - and may choose to bail out - before the data is written to the target table.
    Restrictions
    We have very limited permission to perform DDL e.g. to create new tables/indexes etc.
    We have no permission to use e.g. Oracle external tables, Oracle directories etc.
    We are working with standard Oracle tools i.e. PL/SQL and no DWH tools.
    DBAs are extremely resistant to giving us more disk space.
    We are on Oracle 9iR2, with no immediate prospect of moving to 10g.
    Current approach
    Source data is uploaded via SQL*Loader into static "short fat" tables.
    Some initial key validation is performed on these records.
    Dynamic SQL (plus BULK COLLECT etc) is used to pivot the short-fat data into an intermediate long-thin table, performing the validation on the fly via a combination of including reference values in the dynamic SQL and calling PL/SQL functions inside the dynamic SQL. This means we can pivot+validate the data in one step, and don't have to update the data with its validation status after we've pivoted it.
    This upload+pivot+validate step takes about 1 hour 15 minutes for around 15 million variable values.
    The subsequent "load to target table" step also has to apply substitution rules for certain "special values" or NULLs.
    We do this by BULK collecting the variable values from the intermediate long-thin table, for each valid case in turn, applying the substitution rules within the SQL, and inserting into/updating the target table as appropriate.
    Initially we did this via a SQL MERGE, but this was actually slower than doing an explicit check for existence and switching between INSERT and UPDATE accordingly (yes, that sounds fishy to me too).
    This "load" process takes around 90 minutes for the same 15 million variable values.
    Questions
    Why is it so slow? Our DBAs assure us we have lots of table-space etc, and that the server is plenty powerful enough.
    Any suggestions as to a better approach, given the restrictions we are working under?
    We've looked at Tom Kyte's stuff about creating temporary tables via CTAS, but we have had serious problems with dynamic SQL on this project, so we are very reluctant to introduce more of it unless it's absolutely necessary. In any case, we have serious problems getting permissions to create DB objects - tables, indexes etc - dynamically.
    So any advice would be gratefully received!
    Thanks,
    Chris

    We have 8 "short-fat" tables to hold the source data uploaded from the source file via SQL*Loader (the SQL*Loader step is fast). The data consists simply of strings of characters, which we treat simply as VARCHAR2 for the most part.
    These tables consist essentially of a case key (composite key initially) plus up to 250 data columns. 8*250 = 2000, so we can handle up to 2000 of these variable values. The source data may have 100 any number of variable values in each record, but each record in a given file has the same structure. Each file-load event may have a different set of variables in different locations, so we have to map the short-fat columns COL001 etc to the corresponding variable definition (for validation etc) at runtime.
    CASE_ID VARCHAR2(13)
    COL001 VARCHAR2(10)
    COL250     VARCHAR2(10)
    We do a bit of initial validation in the short-fat tables, setting a surrogate key for each case etc (this is fast), then we pivot+validate this short-fat data column-by-column into a "long-thin" intermediate table, as this is the target format and we need to store the validation results anyway.
    The intermediate table looks similar to this:
    CASE_NUM_ID NUMBER(10) -- surrogate key to identify the parent case more easily
    VARIABLE_ID NUMBER(10) -- PK of variable definition used for validation and in target table
    VARIABLE_VALUE VARCHAR2(10) -- from COL001 etc
    STATUS VARCHAR2(10) -- set during the pivot+validate process above
    The target table looks very similar, but holds cumulative data for many weeks etc:
    CASE_NUM_ID NUMBER(10) -- surrogate key to identify the parent case more easily
    VARIABLE_ID NUMBER(10) -- PK of variable definition used for validation and in target table
    VARIABLE_VALUE VARCHAR2(10)
    We only ever load valid data into the target table.
    Chris

  • Define Tabs and Process Configuration for Template (HAP_TA_CONF)

    I am currently building a new appraisal template for my client and I have configured the process timeline in the development system via the IMG node Define Tabs and Process Configuration for Template (transaction code HAP_TAB_CONF).
    Does anyone know how to transport the tabs and process configuration through the system landscape?
    We are on SAP ECC 6.0 Enhancement Pack 4

    Hi Sushil,
    I have used the report RHMOVE30 as you recommended and it worked perfectly in one run. I did not have to create custom relationships. I simply selected all the Process Item (VH) objects for my appraisal template and ran the report.
    Many thanks for your help. It has saved me having to configure the tabs and process timeline in each client.
    Janet

  • Server for Web hosting and email (only  for 3 domains)

    Hello guys,
    I want to setup mac mini server for my company which will host only 3 web sites and email server for the moment.
    What i got so far,
    01. I have Macmini server
    02. I have static ip address
    03. I have router (vigor 2700) which forward the external ip address (my ip start from 87.197.XXX.XXX) in to my Server using DMZ hosts
    (So i have 192.168.1.20 as my server ip address(local) and when i point 87.197.132.48  from out side (using web browser) i can see Mac server page)
    So can i setup the server using this infomation. Just want to make sure that i allready got everythign to setup the server.
    1st question : Do i have to all information to start this ?
    Please help me with this first, then i will have more  questions latter.
    Thanks in advance 

    Migrate out of 192.168.1.0/24 (and 192.168.0.0/24) as those will cause problems for your eventual use of VPNs to connect into your network.  Most every coffee shop and home network on the planet uses those, and the IP routing inherent in a VPN doesn't appreciate finding the same subnet on both ends of the VPN.  The earlier you get your network out of those particular subnets - and preferably over to a subnet somewhere within the 172.16.0.0/12 or 10.0.0.0/8 blocks - the easier the network migration will be.
    Set up your internal DNS services and your external DNS services (I'm assuming your local DNS server will not be delegated to serve DNS publicly) and particularly establish an A record in (at least) your public DNS referencing your mail server, and then enter that mail server A record entry as the MX mail exchange host for each of the three domains, and set up three virtual domains on the mail server. 
    If your gateway is sufficiently capable, you can have a separate domain or subdomain within your network, and have all your "public" names hosted externally.  You can then use example.com outside, and example.net inside, and any example.com references from the outside or from the inside will resolve to your external static IP address and (if your gateway has the ability to "reflect" traffic to the public static IP) it'll all work most easily.  If your gateway isn't capable, then you can also host the example.com domain internally in your local DNS server, but that means maintaining two parallel sets of DNS entries.  Or you can get a more capable gateway.  (I did not research your gateway.)
    You'll want to read the Apple mail service documentation around alias entries, as there are two ways to do this.  See the stuff around pages 73 and 77 in that document.  Better still, just skim that mail service administration manual; that'll get you an introduction to the whole area and some familiarity with the general terms and concepts, and that manual is fairly readable.
    Then read the TopicDesk Spam Documentation.
    If you don't have IT staff around (some of your time, as well as that of others) to dedicate to this mail server management (including user management, troubleshooting, backups, etc.) hosting mail externally can be a good trade-off.

  • HT5731 How can I download my TV episodes to my Macbook Pro. I bought them from my iPad2, and I looked for them with the same AppleID on iTunes Store and they wont download to my iTunes.

    How can I download my TV episodes to my Macbook Pro. I bought them from my iPad2, and I looked for them with the same AppleID on iTunes Store and they wont download to my iTunes.

    What country are you in ? Depending upon what country that you are in then you might be able to redownload them on your Mac's iTunes via the Purchased link under Quicklinks on the iTunes store homepage
    If you're not in a country where TV programmes can be redownloaded then connect the iPad to your Mac's iTunes and use the File > Devices > Transfer Purchases to copy them over from your iPad

Maybe you are looking for