Enhancment of datasource related query

Hi BW Guru's
I have datasource which having 100 fields and enhanced the same with few more new fields.But i required 60 of the existing as well new fields in BW.How would i select those fields in the datasource and get it the same in to BW.
Please some one could help me in this scenario.
Regards
Prasad

Hi Aryan,
I think you can go to RSA6 and select your data source and bring it to change mode. Then you can simply check the HIDE boxes for those fileds which you don't want to bring to BW. Then save your changes. After that go to RSA3 and check whether your changes have populated there. Once you are done with that just replcate your data source and follow the same process.
Regards,
Jack...

Similar Messages

  • How to enhance the datasource 0SRM_TD_PO using table BBP_PDIHP

    In R3 when any PO is deleted at the header/item level the field LOEKZ is populated in tables EKKO(Header)/EKPO(item level).
    Similarly in SRM when any PO is deleted at the header/item level the field DEL_IND is populated in tables BBP_PDHGP(Header)/BBP_PDIHP(item level). In current situation we have standard datasource 0SRM_TD_PO, this has a field as ITM_DEL_IND which is coming from table BBP_PDIHP(item level) and its always blank and its not getting populated , so we are not able to eliminate the deleted PO's in BW.
    To overcome this, I have a following approach :
    =====================================
    Enhance the datasource 0SRM_TD_PO to include a "ZZDELIND" field and populate this field via CMOD from tables BBP_PDHGP(Header)/BBP_PDIHP(item level). The issue is I'm not able to determine the relation/join/data model between the datasource and the tables. The datasource 0SRM_TD_PO has a PO#(OBJECT_ID) but the tables BBP_PDHGP(Header)/BBP_PDIHP(item level) dont have a PO# instead these tables have a key as "GUID"which is HEXA decimal.
    Can someone please help me out.

    two ways:
    1. call FM bapi_po_getdetail and pass the PO number, you will get all your details
    2. pass PO number to CRMD_ORDERADM_H to get the the GUID and use this GUID on BBP_PDHGP(if version control is active then you need to consider the active version only)
    and for item guid, pass the GUID picked from CRMD_ORDERADM_H to CRMD_ORDERADM_I-HEADER, you will get all the item guids as CRMD_ORDERADM_I.
    this can be passed to BBP_PDIGP.
    there are beautiful views as well joining these tables(BBP_PDVIEW*).. which you can use as well

  • Knowing DataSource, Cube, Query, how to identify Master Data datasource?

    Hi,
    based on a research on FI-FM at http://help.sap.com/saphelp_nw70/helpdata/en/71/d3143cb1b8be00e10000000a114084/frameset.htm
    I saw the flow as: first Query(based on cube 0FIFM_C01), the cube which gets data from 2 datasources (0FI_FM_1  & FI_FM_2) through 2 infosources: 0FI_FM_1 and  FI_FM_2.
    1. What do I need to install on BI to get this going and in which order?
    2. In particular, knowing DataSource, Cube, Query do I figure out the Master Data data I need to get this query functional?
    Once you know of the specific master data datasources what do I need to do to get this query functional?
    Thanks

    Dear Amanda,
    Let's clarify your questions.
    1.Install the cube and related stuff.
    RSA1 -> Business content -> Choose " in data flow before" -> install the Cube
    2. For the master data and related hierarchy.
    RSA1 -> Find the InfoCube -> Right click, find a option such as "display data model". I am not sure about the exactly description because i am not in front of the SAP system.
    you can find all the related InfoObject in the right panel.
    It is good if you check all of those master data and load them daily. But it is a big job and sometimes it is not necessary. You can:
    1) Clarify the business need. to konw which kinds of fields are "Must" for the end user anlaysis and make a list of that.
    2) check the list and most of the master data will load by other moduels. for example, cost center, material, customer ... they are too common that you can expect they are in some process chain and loaded daily basis already.
    3) load the master data in your list and no one load it before, or discuss with the ownder of process chain, if you need change the load frequency such as from weekly to daily.
    Edited by: Brian Yang on Jul 12, 2009 8:54 PM

  • The workbook report no authorization, but related query can show part data

    Dear All,
    We meet an error, The workbook report no authorization, but related query can show part data.
    We have a workbook, it have 1 queries, and when we key in condition to run it, in one field we put on one more condition(Z3XXX = ' WZO'  & ' WZA")(User didn't have authorization for this condition),  and workbook report no authorization.
    Then we checked  authorization check log and it show "You do not have sufficient authorization on Z3XXX = ' WCO'.
    But when we execute the related query with same condition. the query can open and show WZA's data.
    I'd like to know the reason.
    thanks & Best Regards,
    Kent Yu

    Dear All,
    We meet an error, The workbook report no authorization, but related query can show part data.
    We have a workbook, it have 1 queries, and when we key in condition to run it, in one field we put on one more condition(Z3XXX = ' WZO'  & ' WZA")(User didn't have authorization for this condition),  and workbook report no authorization.
    Then we checked  authorization check log and it show "You do not have sufficient authorization on Z3XXX = ' WCO'.
    But when we execute the related query with same condition. the query can open and show WZA's data.
    I'd like to know the reason.
    thanks & Best Regards,
    Kent Yu

  • Multi-part related query reports

    http://download-east.oracle.com/docs/html/B10310_01/orbr_concepts1.htm#1012365
    Can someone provide me better info about 'Multi-part related query reports '

    Hi,
    yes you can have multiple query in your report. So you have multiple groups in your data model. Then you must put multiple repeating frames in your layout in order to extract data from your queries, selecting the desired group in the source property of your repeating frame.
    Bye
    Raffy

  • Enhancing Text Datasource

    Hi
    I want to enhance 0STOR_LOC_TEXT datasource with 0COUNTRY and country key. but, can i use them as a navigational attribute in the report, after adding into 0STOR_LOC_TEXT ?
    Regards
    Anuja

    Hi Anuja,
    For making any infoobject as navigational you need to do following setting
    1)Infoobject should be navigational as master data infoobject.
    2)infoobject should be navigational at data target level(it can be your DSO,Cube,multiprovider).
    Please do above setting at your data model then only you can use that infoobject as navigational at report level.
    Hope it will helpful to you.
    Regards,
    Ganesh Bothe

  • Steps to enhance a datasource from SAP CRM

    Dear Forum,
    Could you please provide me with some document to enhance a datasource from SAP CRM. Is it any different from enhancement of datasource from ECC datasources,
    Thanks and regards,

    Hi,
    please look this:
    Business Intelligence How-to Guides for SAP NetWeaver 2004 [original link is broken]
    or here: How-to Guides for SAP NetWeaver Business Warehouse  7.0 [original link is broken]
    Creating a datasource for generic extraction
    regards
    CSM Reddy

  • Delta in enhanced 0HR_PT_1 DataSource

    Hi gurus,
    I have a problem with enhanced 0HR_PT_1 DataSource.
    There are some fields added to the extractor structure and one of these fields (ZNIEOB) should be the base for delta queue.
    Unfortunatelly the delta mechanism doesn't work as it should - the changes in ZNIEOB field don't generate delta records.
    How should I modify the extractor to make the delta work?

    Hi,
    Please check if pseudo delta helps:
    dummy delta/pseudo delta
    -Vikram

  • Steps to Enhance Custom DataSource

    Hello All,
    I have one custom datasource ZZ_PROD. From which I am extracting data and populatd in a ODS and then in to Cube in BI 7.0
    Now I have to enhance it and add about more 20 fields from two different ECC tables.
    These two tables are custom tables for a particular departmens.
    Can somebody explain me the steps  how I can accomplish this?
    Thanks in Advance
    Kumar M.

    I would recommend to change the function Module or the program it generates the ZZ_PROD.
    First Add the fields to the extract Structure.
    Unhide the newly added fields in the data source.
    Write the logic to grab the data from the two tables and assign it to proper extract structure fields.
    We normally enhance the datasource using User Exit since we don't want to make changes to SAP Delivered programs that would be overwritten during upgrades.
    Hope that helps.
    thanks.
    Wond

  • BI datasources related to country specific found missing?

    hi everyone,
    i am not aware whether to post here or not...
    i am using IDES system of R3 ECC 6.0  and BI 7.0 but not aware of components in it..
    i am looking for datasources related to country specific(ex: 0ID_BR_NF_1 etc) in rsa5 tcode in R3.datasources seems missing..
    can anyone suggest what component is to be added for ECC.and BI
    and give me some idea about ADDON and SP(service pack).
    i had gone through tcode SAINT. please priscribe what addon should be added

    Hi JP,
    I think I will need to explain a little bit more/different.
    First of all, thanks for your help, I will note these exit's for future issues, who knows?
    The userexits you mentioned are helpful when the proposal was created and needs to be copied.
    But in my case it is a bit different. I created new keys/conditions for packages (key including route) plus corresponding rules using POP1/POF1. And I want these rules to be used when creating packing proposal with button "automated packing". But my new rule with new key has not been used. Only rules with the "old"/standard key were taken into consideration.
    So I assume, that for automated packing I will need to read also the "new" key in any exit. And this exit should be anywhere while the proposal will be created in the order.
    But another possibility is, that SAP checks the condition type hardcoded.
    Any idea?
    Best regards,
    Gerlinde

  • How can I related Query name with Query elements?

    Hello gurus!
    I am working with BI7, content technical, well I need to find out a solution where I can related Query name with query elements (KYF, CKY, Restricted Key Figures, var, etc)
    In master data 0tctquery, exist the query element, but, I can not related with object are from what query...
    Do you know how I can do it??? some solution???
    Best Regards

    "it is impossible to relate both tables" ???
    hmmm... these are exactly the tables the Query Designer uses to "build" up the query... so trust me, they do relate
    now to get this information in an InfoCube is a totally different thing... how do you expect to report on this if you should get it in an InfoCube? let's take a simple example... assume you have a query (QUERY_1) which looks like:
                                      Sales  Sales
    Sales Org Customer    Qty      Value
    1000         ABC           10 PC  500 €
    1000         DEF             5 PC  250 €
    in your Cube you would get:
    QueryName  QueryElement
    QUERY_1    0SALESORG
    QUERY_1    0CUSTOMER
    QUERY_1    ZSLSQTY
    QUERY_1    ZSLSVAL
    if you have free characteristics, you'll add more lines... same for variables, CKFs, RKFs, conditions, ...
    I guess it would be better to create a little ABAP report using the above tables (oh, you'll be needing a lot more tables if you wish more detail)

  • Converting Relational query to XML

    set serveroutput on
    This is the code of a query to convert a relational query to xml. This code runs fine when I place the where condition at the end. e.g. where h.hospital_id=1. however when i place no such condition it begins returning errors such as
    set serveroutput on
    set long 10000000
    set linesize 50000
    set lines 20000
    set pagesize 0
    set pages 5
    set feedback off
    set head off
    SET TRIMSPOOL ON
    set long 200000000
    set linesize 32000
    set pagesize 0
    set pages 0
    set feedback off
    set head off
    Select XMLELEMENT("have:HospitalStatus",XMLAttributes('urn:oasis:names:tc:emergency:EDXL:HAVE:1.0 edxl-have_cs01.xsd' AS
                                      "xsi:schemaLocation",
                                              'urn:oasis:names:tc:emergency:EDXL:HAVE:1.0' AS
                                      "xmlns:have",
                                    'http://www.georss.org/georss' AS
                                      "xmlns:n3",
                                    'http://www.opengis.net/gml' AS
                                      "xmlns:gml",
                                    'http://www.w3.org/2001/XMLSchema-instance' AS
                                      "xmlns:xsi",
                                    'urn:oasis:names:tc:ciq:xal:3' AS
                                      "xmlns:xal",
                                    'urn:oasis:names:tc:ciq:xnl:3' AS
                                      "xmlns:xnl",
                                    'urn:oasis:names:tc:ciq:xpil:3' AS
                                      "xmlns:xpil"                                        
         xmlagg(                    
         xmlelement("have:Hospital",
         xmlforest
              (     h.hospital_id,          
                   xmlforest(
                   xmlforest(
                   h.name,               
                   h.info,          
                   h.addresses     ,          
                   h.contact_numbers) as "OrganizationInformation",     
                   xmlforest(
                   h.geo_location) as "OrganizationGeoLocation",
                   h.comments) as "Organization"
                   xmlelement("have:EmergencyDepartmentStatus",
                   (select xmlagg(
                   xmlforest(
                             eds.date_updated,
                   xmlforest(
                             eds.ems_traffic_status,
                             eds.ems_traffic_reason) as "EMSTraffic",          
                   xmlforest(
                             tcl.url,
                             tc.date_updated,
                             tc.code_value,
                             tc.ems_capacity
                        ) as "EMSCapacity",     
                   xmlforest(
                             tcl.url,
                             tc.date_updated,
                             tc.code_value,
                             tc.ems_census) as "EMSCensus",
                             xmlforest(
                             eds.ems_amb_offload_status,
                             eds.ems_amb_offload_min) as "EMSAmbulanceStatus",
                             xmlforest(
                             eds.ems_air_offload_status,
                             eds.ems_air_offload_min )as "EMSAirTransportStatus"
                   from EmergencyDepartmentStatus eds,TriageCode tc,TriageCodeListURN tcl
                   where eds.hospital_id = h.hospital_id and tc.hospital_id = h.hospital_id and tcl.hospital_id = h.hospital_id)),     
                             xmlelement("have:HospitalBedCapacityStatus",
                   (select xmlagg(xmlelement("have:BedCapacity",
                   xmlforest(
                             mb.date_updated,
                             mb.bed_no,
                             mb.bed_type,
                             xmlforest(
                             mb.capacity_status,
                             mb.available_count,
                             mb.baseline_count,
                             mb.additional_24hr,
                             mb.additional_72hr) as "Capacity",
                             mb.comments
                   from MainBeds mb
                   where mb.hospital_id = h.hospital_id and mb.bed_no NOT IN(select mb.bed_no from MainBeds mb, SubBeds sb where mb.bed_no=sb.bed_no )
                   (select xmlagg(xmlelement("have:BedCapacity",
                   xmlforest(
                             mb.date_updated,
                             mb.bed_no ,
                             mb.bed_type,
                             xmlforest(
                             mb.capacity_status
                                                 ) as "Capacity",
                             mb.comments
                                  ),(select xmlagg(
                   xmlforest(
                                  sb.date_updated,
                                  sb.bed_no,
                                  sb.sub_bed_no,
                                  xmlforest(
                                  sb.capacity_status,
                                  sb.available_count,
                                  sb.baseline_count,
                                  sb.additional_24hr,
                                  sb.additional_72hr) as "Capacity",
                                  sb.comments
                                  )from SubBeds sb  where mb.bed_no = sb.bed_no and sb.bed_no=h.hospital_id )
                   from MainBeds mb
                   where mb.hospital_id = h.hospital_id and mb.bed_no in (select unique mb.bed_no from MainBeds md, SubBeds sb where sb.bed_no=mb.bed_no))
                   xmlelement("have:ServiceCoverageStatus",
                   (select xmlagg(
                   xmlforest(
                             scs.date_updated,
                             scs.burn,
                             xmlelement("CardiologySubType",
                                  xmlforest(scs.cardiology_invasive,
                                  scs.cardiology_non_invasssive)) as "CardiologyIndicator",
                             scs.dialysis,
                             scs.emergency_dept,
                             scs.hyperbaric_chamber,
                             scs.infectious_diseases,
                             scs.neonatology,
                             xmlforest(xmlforest(
                                  scs.neurology_invassive,
                                  scs.neurology_non_invasssive) as "NeurologySubType") as "NeurologyIndicator",
                             xmlforest(xmlforest(
                                  scs.obgyn_with_labor,          
                                  scs.obgyn_without_labor) as "OBGYNSubType")as "OBGYNIndicator",
                             scs.opthalmology,
                             scs.orthopedic,
                             scs.pediatrics,
                             xmlforest(
                                  xmlforest(scs.psychiatry_adult_general,
                                  scs.psychiatry_pediatric) as"PsychiatrySubType")as "PsychiatryIndicator",
                             xmlforest(
                                  xmlforest(
                                  sst.date_updated,
                                  sst.general_ward,
                                  sst.adult_general,
                                  sst.pediatrics,
                                  sst.orthopedics,
                                  sst.nuero_surgery,
                                  sst.facial,
                                  sst.cardio_thoracic,
                                  sst.hand,
                                  sst.reimplantation,
                                  sst.spinal,
                                  sst.vascular,
                                  sst.anesthesia) as "SurgerySubTypes") as "SurgeryIndicator",
                             xmlforest(xmlforest(
                             scs.ambulance_services,
                             scs.air_transport_services) as "TransportServicesSubType") as "TransportServicesIndicator",
                             xmlforest(xmlforest(
                             scs.trauma_service,
                             scs.trauma_service_level) as "TraumaCenterServicesSubType")as "TraumaCenterServicesIndicator"
                   from ServiceCoverageStatus scs, SurgerySubTypes sst
                   where scs.hospital_id = h.hospital_id and sst.hospital_id = h.hospital_id)),
                   xmlelement("have:HospitalFacilityStatus",
                   (select xmlagg(     
                   xmlforest(
                             hfs.date_updated,               
                             hfs.hospital_eoc_status,
                             hfs.hospital_eoc_plan,          
                             hfs.clinical_status,
                             xmlforest(
                                  dc.date_updated,
                                  dc.decon_capacity_status,
                                  dc.amb_patients_capacity,
                                  dc.non_amb_patients_capacity) as "DeconCapacity",
                             xmlforest(
                             mc.date_updated,
                             mc.morgue_capacity_status,
                             mc.morgue_capacity_units) as "MorgueCapacity",
                             hfs.facility_status,               
                             hfs.security_status,     
                             xmlforest(
                                  ah.date_update,
                                  ah.admissions,
                                  ah.discharges,
                                  ah.deaths) as "Activity24hr",
                             hfs.comments,
                             hfs.clinical_operations,     
                             hfs.resources_information
                   from HospitalFacilityStatus hfs,Activity24hr ah,MorgueCapacity mc,DeconCapacity dc
                   where hfs.hospital_id = h.hospital_id and ah.hospital_id = h.hospital_id and mc.hospital_id = h.hospital_id
                   and dc.hospital_id = h.hospital_id)),     
                   xmlelement("have:HospitalResourceStatus",
                   (select xmlagg(
                   xmlforest(
                             hr.date_updated,
                             hr.staffing,
                             hr.facility_operations,
                             hr.clinical_operations,     
                             hr.comments
                   from HospitalResourceStatus hr
                   where hr.hospital_id = h.hospital_id)),
              xmlforest(h.last_update_time) as "UpdateTime"
              ))).extract('/*')
    from hospital h;__________________________________________________________________________________________________________
    ERROR:
    ORA-31011: XML parsing failed
    ORA-19202: Error occurred in XML processing
    LPX-00225: end-element tag "HOSPITAL_EOC_STATUS" does not match start-element tag "have:HospitalStatus"
    Error at line 1
    ORA-06512: at "SYS.XMLTYPE", line 111
    Edited by: 872489 on 14-Jul-2011 03:59
    Edited by: 872489 on Jul 14, 2011 9:11 AM
    Apologies for lack of formatting earlier

    Please repost this and use the tag with some decent formatting.
    See the FAQ Posting Guidelines on the main page of this XMLDB OTN Forum                                                                                                                                                                                                                                                                                               

  • Enhancement BC Datasource for Master Date  - no data for new fields

    Hello,
    I have a problem. Maybe someone can provide a check list.
    I made an enhancement for datasource 0COORDER_ATTR via SAP Exit. The problem is that i do not get data in BW for the new fields but no error occurs. the extractor checker delivers the new data.
    the following points i already checked:
    - datasource replicated and infosource activated
    - "Hide field" and "Field only known in customer exit" are not checked for the new fields

    Hi Stefan,
    I guess you need to maintain your transfer structure. Edit your transfer rules and click on tab 'datasource/transfer structure'. Move the new fields from the right table (datasource) to the left table (transfer structure). Click on tab transfer rules and map the new field. Activate the transfer rules again.
    regards
    Siggi

  • How does an enhancement of datasource affect my production data

    Hello All,
    We have enhanced a datasource 2LIS_11_VAITM, I want to know what will happen to my existing data in the production after i move the enanced datasource to production???
    will i need to clear the existing data in production after moving the same to another dummy cube O r will just running the deltas bring in the data for the enhanced fields as well without disturbing the existing data...
    Please respond. The issue is very critical..

    Hi,
    For all the historical data that already exists in your BI system, the data for the enhanced fields will not be populated but deltas would work absolutely fine and for all the new transactional changes, these fields will be filled. If you want to populate the data for historical data as well, you will have to carry out fill up of setup tables and reload the data to BI by initialising the data. If the data is being fed to DSO, I dont think that there is any need for deleting the data, as it would get overwritten by the new initialisation load.
    As such, please note that deltas will not get triggered if the only changes to Sales Order are for enhanced fields. Deltas will only be captured for the fields that are part of main communication structure. The motive for adding the enhaned fields is only to bring information for these fields.
    Edited by: Rahul K Rai on Sep 14, 2010 11:26 AM

  • SSAS using datasource at query time

    I have a set of reports that run against the cube.  The cube has proactive caching enabled.  If the datasource within the SSAS database has invalid credentials, queries against the cube will not run.  Note: SSAS is setup to run as network
    service rather than a domain account (temporary situation).
    My understanding of MOLAP is that the data is stored within the cube an no access to the source (sql server datamart) is needed for the user to run queries against the cube.  Now this is proving to not be true... leaving me feeling like I don't know
    what is going on at all.  After 7 years using SSAS I'm finding out that the user's queries don't actually run against the cube, they require that ssas be able to connect to its source data to satisfy the query???!?!?!??!?
    I really really don't understand this, but I have tested repeatedly and when I set the datasource to invalid credentials, report queries won't run or just return 0's.  Earlier we were getting 08001 "can't connect to server" messages, now I'm
    just getting all 0's on the reports when the datasource credentials are invalid.
    Just to make this 100% clear, I am talking about the SSAS datasource within the SSAS database, not the SSRS datasource that points to the cube.  Has anyone ever seen this before?  To make it more confusing, now it seems to be working in excel regardless
    but SSRS still needs the valid datasource.
    Thanks,
    Ken

    Hi Ken,
    The data source credential of SSAS cube you specified in SSMS was use to retireve data from relational database. For example, we will need to update cube data or some MDX queries need to get data from relational DB. The Analysis Services querying architecture
    provides several components that work together to efficiently retrieve and evaluate data, please see:
    When we query a cube, the query processor breaks the query into subcube requests for the storage engine. For each subcube request, the storage engine first attempts to retrieve data from the storage engine cache. If no data is available in the cache, it
    attempts to retrieve data from an aggregation. If no aggregation is present, it must retrieve the data from the fact data from a measure group’s partition data.
    In your case, I recommend you use SQL Server Profiler to capture SSAS and Database Engine events for further investigation. If the SSRS reports need to get data from relational database, we will get detail information in this scenario.  For more information
    regardig SQL Server Profiler, please see:
    http://technet.microsoft.com/en-us/library/ms181091.aspx
    If you have any feedback on our support, please click
    here.
    Regards,
    Elvis Long
    TechNet Community Support

Maybe you are looking for