Update Rule 8ZCDSDS03 and 80CDS_DS07 for data mart

Hello Guru´s
My proble is this
8ZCDSDS03 Update rules to Data marts for cube 0SD_C03
80CDS_DS07 Update rules to Data marts for cube 0SD_C11
are in gray i don´t know why. do you have and explanation or a solution for that.
whit best regards
agt

Hi Antonio,
Any time you make a change to a cube, check the update rules feeding it and activate them. Also check if this InfoCube is supplying data to other data targets and check the 8* update rules and activate if required.
Right now you can double click the grey rules and activate them and transport the active rules (if required)
Hope this helps...

Similar Messages

  • [svn:fx-trunk] 8786: Minor updates from Sujata and Min for fxgutils.

    Revision: 8786
    Author:   [email protected]
    Date:     2009-07-23 21:10:35 -0700 (Thu, 23 Jul 2009)
    Log Message:
    Minor updates from Sujata and Min for fxgutils. Addresses an issue where FXG files with a matrix display differently after running through FXG to FXG. Also includes a fix to stop processing of unknown fxg tags in skipped sections.
    QE: No
    Doc: No
    Checkintests: Pass
    Reviewer: Min and Sujata
    Modified Paths:
        flex/sdk/trunk/modules/fxgutils/src/java/com/adobe/internal/fxg/dom/filters/AbstractFilte rNode.java
        flex/sdk/trunk/modules/fxgutils/src/java/com/adobe/internal/fxg/dom/transforms/MatrixNode .java
        flex/sdk/trunk/modules/fxgutils/src/java/com/adobe/internal/fxg/sax/FXGSAXParser.java
        flex/sdk/trunk/modules/fxgutils/src/java/com/adobe/internal/fxg/sax/FXGSAXScanner.java

    This is a duplicate post.  This should be locked to avoid further confusion.
    My CSS wont change
    Nancy O.

  • Copy config and payroll rules & shemas and 2 years data into ECC6 from 4.7

    Hi ,
    we need to copy config and payroll rules & shemas and 2 years data into ECC6 from 4.7 enterprise .
    Please let me now How to do it ?
    I will give points for suitable answers pl
    Thanks,
    Sunitha

    Hi Sonu Gupta,
    Can please explain bit more ...
    Config and Data both i wanted to move 4.7 sytem to ECC6 system . both are diffrent systems .
    what is this peace list ?How to create ?
    in case of Config how to trsnfer .
    Please Help Me ...
    Sunitha
    Edited by: Sunitha Reddy on Apr 10, 2008 3:34 PM

  • Start routine in Transfer Rules for Data Mart update

    Hi,
    Is it possible to create a start routine in a transfer rule of a Datamart update (from ODS to Cube)?
    Thxs.
    LauQ

    Roberto,
    If I read well the content of note 499404 - referred to in 630854 - I conclude that it is recommended to use update rules rather than transfer rules in the case of ODS objects when one wishes to modify data records up to the eventual data target.
    Do you agree?
    Thxs.
    LauQ

  • Table Onwers and Users Best Practice for Data Marts

    2 Questions:
    (1)We are developing multiple data marts that share the same Instance. We want to deny access to the users when tables are being updated. We have one generic user (BI_USER) with read access through one of the popular BI Tools. The current (first) data mart we denied access by revoking the privilege to the BI_USER, however going forward with other data marts the tables will get updated on a different schedule and we do not want to deny access to all the data marts. What is the best approach?
    (2) What is the best Methodology for table ownership of tables in different data marts that share tables across marts? Can we create one generic ETL_USER to update tables with different owners?
    Thanx,
    Jim Masterson

    If you have to go with generic logins, I would at least have separate generic logins for each data mart.
    Ideally, data loads should be transactional (or nearly transactional), so you don't have to revoke access ever. One of the easier tricks to accomplish this is to load data into a shadow table and then rename the existing table and the shadow table. If you can move the data from the shadow table to the real table in a single transaction, though, that's even better from an availability standpoint.
    If you do have to revoke table access, you would generally want to revoke SELECT access to the particular object from a role while the object is being modified. If this role is then assigned to all the Oracle user accounts, everyone will be prevented from viewing the table. Of course, in this scenario, you would have to teach your users that "table not found" means that the table is being refreshed, which is why the zero downtime approach makes sense.
    You can have generic users that have UPDATE access on a large variety of tables. I would suggest, though, that you have individual user logins to the database and use roles to grant whatever ad-hoc privileges users need. I would then create one account per data mart, with perhaps one additional account for the truely generic tables, that own each data mart's objects. Those users would then grant different roles different database privileges, and you would then grant those different roles to different users. That way, Sue in accounting can have SELECT access to portions of one data mart and UPDATE access to another data mart without granting her every privilege under the sun. My hunch is that most users should not be logging in to, let alone modifying, all the data marts, so their privileges should reflect that.
    Justin
    Distributed Database Consulting, Inc.
    http://www.ddbcinc.com/askDDBC

  • Update rule not working with high data load.

    Hi all,
    i have a problem with a update rule: it's an update loop in a dso, in the start routine i do 3 select for all entries in data-package on the active table of another structure; then i read those table to update some values in my update rule.
    I did some test and it seemed to work well (i tryed for example just for a plant) but when i launched it for all the records in the dso the result was differente and, for the same plant, the values where not updated correctly (they were blank).
    Now the routine is really the same so it sound strange to me that launching the infopackage without filters i can't get the same correct result of my previous test but i was wondering what could be the reason of this error...
    Anyone can help?
    The start routine is this:
      REFRESH i_tab.
      SELECT field1 field2 field3
        INTO TABLE i_tab
        FROM target_dso
        FOR ALL ENTRIES IN DATA_PACKAGE
        WHERE deliv_numb = DATA_PACKAGE-deliv_numb
          AND deliv_item = DATA_PACKAGE-deliv_item
          AND act_gi_dte <> l_blank_date.
    then i read this table in the other routines...

    It is hard to say.  What does the update rule look like?
    after the read statement, you could check the return code.  If it is not zero, go into an infinite loop, and debug it via SM37.
    read table....
    IF sy-subrc <> 0.
      WHILE 1 = 1.
        "debug in SM37.
      ENDWHILE.
    ENDIF.

  • BW 3.5 Update rule routine and start routine convert to BI 7.0 Endroutine.

    I have bw 3.5 update routine and update rules start routine( r/3 to ODS). i need to replicate that routine into BI 7.0 endroutine with the same logic with some minor changes(DSO to DSO).
    IN BW 3.5 the data is getting from r/3 where as in BI7.0 the data fetching from DSO itself but the logic is same as bw 3.5.
    following is the start routine:
    DATA: G_FISCPER_TO TYPE /BI0/OIFISCPER,
          G_CALMONTH   LIKE /BI0/PCALMONTH-CALMONTH,
          G_MONAT      TYPE /BIC/OIYRDFMONAT.
    ABORT = 0.
    ------ Globale Variable füllen --------------------------------------*
      CALL FUNCTION 'Y_RDF_FISCPER_FOR_INFOPACKAGE'
        EXPORTING
          I_ISOURCE    = '0FI_AA_001' (*** this is infosurce getting from r/3, in 7.0 this one is DSO 'B")
        IMPORTING
          E_FISCPER_TO = G_FISCPER_TO
          E_MONAT      = G_MONAT
        EXCEPTIONS
          OTHERS       = 1.
      IF SY-SUBRC = 0.
        CALL FUNCTION 'Y_RDE_FISCPER_TO_CALMONTH'
          EXPORTING
            I_PERIV    = 'K4'
            I_BDATJ    = G_FISCPER_TO(4)
            I_POPER    = G_FISCPER_TO+4(3)
          IMPORTING
            E_CALMONTH = G_CALMONTH
          EXCEPTIONS
            OTHERS     = 1.
        IF SY-SUBRC NE 0 OR G_CALMONTH IS INITIAL.
          CALL FUNCTION 'Y_RDN_MONITOR_SET'
            EXPORTING
              I_MSGID   = 'YRDFBW'
              I_MSGNO   = '027'
              I_MSGV1   = '0FI_AA_001'
              I_MSGV2   = 'YRDF_FIAA_PERIOD'
            TABLES
              T_MONITOR = MONITOR.
          ABORT = 1.
        ENDIF.
      ELSE.
        CALL FUNCTION 'Y_RDN_MONITOR_SET'
          EXPORTING
            I_MSGID   = 'YRDFBW'
            I_MSGNO   = '027'
            I_MSGV1   = '0FI_AA_001'
            I_MSGV2   = 'YRDF_FIAA_PERIOD'
          TABLES
            T_MONITOR = MONITOR.
        ABORT = 1.
      ENDIF.
    Update routine:
    RETURNCODE = 0.
      ABORT = 0.
      CLEAR RESULT.
      IF NOT COMM_STRUCTURE-ASSET_MAIN BETWEEN '000000200000' AND
                                               '000000299999'.
        CALL FUNCTION 'Y_RDF_GET_CCTR_0ASSET'
          EXPORTING
            I_COMP_CODE  = COMM_STRUCTURE-COMP_CODE
            I_ASSET      = COMM_STRUCTURE-ASSET
            I_ASSET_MAIN = COMM_STRUCTURE-ASSET_MAIN
          IMPORTING
            E_COSTCENTER = RESULT
          EXCEPTIONS
            OTHERS       = 1.
        IF SY-SUBRC NE 0 OR RESULT IS INITIAL.
          CALL FUNCTION 'Y_RDN_MONITOR_SET'
            EXPORTING
              I_MSGID   = 'YRDFBW'
              I_MSGTY   = 'E'
              I_MSGNO   = '061'
              I_MSGV1   = COMM_STRUCTURE-COMP_CODE
              I_MSGV2   = COMM_STRUCTURE-ASSET
              I_MSGV3   = COMM_STRUCTURE-ASSET_MAIN
            TABLES
              T_MONITOR = MONITOR.
          RETURNCODE = 1.
        ENDIF.
      ENDIF.
    How can i write this in BI7.0?
    Thanks

    Hi ,
    solved on my own. Thankyou very much..
    thanks & regards,
    M.S
    Edited by: M.S on Oct 27, 2009 6:57 AM
    Edited by: M.S on Oct 27, 2009 6:59 AM

  • Update Rule error while loading the data

    Hello All,
    I tried to do the init load for the datasource 2LIS_02_VAKON (Sales Document Condition), But it error's out in the Update rule saying:
    Short text: Error when generating the update program
    Long Text:  Diagnosis
         An error occurred during program generation for InfoSource 2LIS_11_VAKON
         and InfoProvider 0SD_C15 . This may be extrapolated to incorrect update
         rules.
    Procedure
         Check and correct your update rules and then generate the update program
         again. You can find help on the error in the error log.
    The update rule looks very fine. I even deleted it and re created, but no success..... Please give me your thoughts abt it..... Any suggestions will be great for me.....

    Hello Gokul,
    I could'nt see any analyse button in RSMON. This is the only error message I could get from the Process monitor:
    Short text: Error when generating the update program
    Long Text: Diagnosis
    An error occurred during program generation for InfoSource 2LIS_11_VAKON
    and InfoProvider 0SD_C15 . This may be extrapolated to incorrect update
    rules.
    Procedure
    Check and correct your update rules and then generate the update program
    again. You can find help on the error in the error log.
    Do you expect something else ? Also I went through your OSS note, they are asking to apply support pack 18. Is this a serious issue to apply SP? Because in Quality and Prod it loads fine......
    Please let me know your idea

  • How to add in ODBC connection and configuration for data connection

    Hello expert,
    on my computor, there is a oracle client for oracle 10g. now there is new oracle 11g installed , I want to create a ODBC connection to this new database. but when I create ODBC , I can't find server type for oracle 11g , there only has driver for oracle 10g. will you please tell me how to create ODBC for this new DB? will you please tell me where is the configuration file for data source on my computor?
    Many Thanks,

    918440 wrote:
    Hello expert,
    on my computor, there is a oracle client for oracle 10g. now there is new oracle 11g installed , I want to create a ODBC connection to this new database. but when I create ODBC , I can't find server type for oracle 11g , there only has driver for oracle 10g. will you please tell me how to create ODBC for this new DB? will you please tell me where is the configuration file for data source on my computor?
    Many Thanks,So, on this machine you have
    - Oracle 10g client (only), and in a seperate ORACLE_HOME
    - Oracle 11g rdbms
    First, the 10g client will connect quite well to the 11g database
    Second, the 11g database includes the basic client admin software, but the odbc drivers are not installed by default. You have to run OUI and choose 'custom' installation, then select 'windows components' from the list of available components.
    I've never wanted to install the odbc driver as part of the database installation. Only on remote client installations.
    What are you really trying to achieve?

  • Vlan for voice and vlan for data with diferent ISP best choice of config??

    Hello everyone,
    Im, Oscar
    At our company we have a redundant ISP connection to two separate ISP's.
    We are also using VoIP on the network.
    Currently one ISP connection is used primarily and the second on is just used as a backup.
    I was wondering if it is possible to use the first connection primarily for data traffic and the second connection for VoIP traffic?
    We use different VLAN's for voice and data.
    Any help would be appreciated.

    Disclaimer
    The Author of this posting offers the information contained within this posting without consideration and with the reader's understanding that there's no implied or expressed suitability or fitness for any purpose. Information provided is for informational purposes only and should not be construed as rendering professional advice of any kind. Usage of this posting's information is solely at reader's own risk.
    Liability Disclaimer
    In no event shall Author be liable for any damages whatsoever (including, without limitation, damages for loss of use, data or profit) arising out of the use or inability to use the posting's information even if Author has been advised of the possibility of such damage.
    Posting
    Yes, for egress.  Ingress is "it depends".
    You could also consider using both links for both kinds of traffic.

  • How to acheive between and condition for dates in prompts

    Hi all,
    can any one tell me how to achieve between and condtion in obiee 11g
    i used like this,
    i in the filter i put active date is between and put two presentation variables(pre_var1 and Pre_var2) and in prompts
    is used to one prompt active date is between when i am passing the same date the result is not getting. the prompt that i used is set a presentation variable(Pre_var1) is it correct?
    can any one help in this

    HI,
              What i got from ur ques is that u want single record for every customer havin oldest start date and most future end date. If it is so then in ur update routine from source ODS to Target ODS try this code :
    Sort data_package by customer startdate ascending enddate descending.
    delete adjacent duplicates from data_package comparing customer.
    after sorting u will have all records of a customer together with oldest start date and most future end date as first record. Delete will remove rest of the records for a customer from data_package.
    Regards:
    Jitendra
    Edited by: Jitendra Gupta on Jul 11, 2009 12:53 PM

  • Updated Direct 7mode And CDOT Script Data Sources

    Here are updated versions of my direct data sources I put together (communities doesn't seem to let me edit the existing article) - the 20150703 update adds support for WFA 3.0 and provides much more complete cDOT (cm_storage), including cDOT 8.3 features. The 20150728 update fixes storage pool support. The 20150802 update adds support for WFA 3.1RC1 and fixes failover_group support in WFA 3.0. Not everyone has pre-existing OCUM and installing and configuring OCUM for both 7mode and cDOT is not necessarily a small task. These datasource types populate the standard storage and cm_storage schemas which can then be used as normal with many workflows. The WFA 3.0 cDOT direct data source is almost complete. It now populates the following: disk, disk_aggregate, efficiency_policy. It includes cDOT 8.3 information added in WFA 3.0: broadcast_domain, failover_group, ipspace, storage_pool*. There may be some minor gaps in some of the data - please let me know if you identify anything. (Note: The WFA 2.2* cDOT data source is missing all of these items). The WFA 3.1RC1 cDOT dats source adds support for cluster_peer, vserver_peer and SVM DR, and all new fields added in 3.1RC1 are populated. It also populates cifs domain information. The 7-mode source is less complete. It contains: array, vfiler, aggregate, interface, volume, vsm, qtree, lun. It is missing: array_license, cifs_share, cifs_share_acl, dataset, disk, igroup, igroup_initiator, lunmap, nfs_export, object_comment, quota, snapshot, snapvault, user_quota. Configuration is simple:Provide credentials for the arrays/clusters in WFA credentials.Add datasource - select version appropriate to your WFA version (the 7-mode source works with 2.2-3.0) which the cDOT one is different for 2.2 and 3.0.Set the Hostname of the datasource to a comma separated list of array or cluster admin names or addresses, or set up one datasource per array/cluster.Configure the interval to something reasonable (say 15 or 30 minutes)Increase the timeout for larger environments (in my testing, collection takes 15-90s per array/cluster - it may take longer when there are large numbers of objects or if there is significant latency between the WFA system and the array/cluster) There are some flags that can be passed in (comma separated) through the data source's "User name" field:strict - fail collection if any error occurs (default if there is only one array/cluster)nostrict - don't fail connection if an error occurs (for a single array/cluster datasource)debug - log more informationlog=<logpath> - where to log (defaults to ..\..\log\direct_<schema>.log - typically C:\Program Files\NetApp\WFA\jboss\standalone\log\)timeout=<timeout_in_milliseconds> - API timeout (defaults to 180000)The Password, Database and Port fields are unused. Reservations, etc, should work as normal. This has been developed against WFA 2.2RC1, 2.2, 2.2.1, 3.0RC1 and 3.0 and will likely require updates to work with future versions. Upgrades from previous versions should be seamless. Changes from previous version:Added WFA 3.0 schema supportAccomodate changed mysql behavior in WFA 3.0Single array/cluster is now automatically strictAdded nostrict parameter7-mode direct data source handles pre 7.3.3 systems7-mode direct data source doesn't fail if vfiler stopped/inconsistentFix for storage_pools in 3.0 cDOT data source (20150728)Fix for broadcast_domain (3.0/3.1RC1)Added support for WFA 3.1RC1, populates all new schema tables and fields.No longer imports into WFA 2.2RC1

    Hi Richard, The dar file is directly attached, not wrapped in a zip as the previous communities site used to do. You should be able to directly import it into WFA. Note that some browsers may save the file with a zip extension (I've seen IE do this - dar files are just zip files anyway), so you may need to rename the file to .dar first. Regards,Tim

  • Is can we develop MATCH btwn VBEP and VBFA, for data pulling?

    Hi Experts,
    I got a requirement, in which I hv to develop a report for the Open quantities for sales order items. fine.
    I ssue is that, in the report I hv to show at Schedule line level i.e. VBEP.
    So, by pulling the records from VBEP depending on the seldction criteria, I hv to pull the MATCHING record from VBFA for Post Goods Issue.
    Then, I hv to implememnt the follwoing formula,
    Open Quantity = VBEP, say Schedule line 1 quantity -(mnus) first saved corresponding Order, Item Quantitity-RFMNG in VBFA for PGI type
    As we know, VBEP is at Schedule line lvel, where as VBFA is ITEM level!!
    So, pls. let me know that,
    1 - Is really, we can develop the MATCH btwn VBEP and VBFA tbls, I mean, always, the data will be correctly pulled?
    2 - If so, Is the folllwoing statement is OK?
    LOOP AT IT_VBFA WHERE VBELV = IT_VBEP-VBELN and POSNV = IT_VBEP-POSNR and ERDAT LE IT_VBEP-MBDAT
    3 - Any other good statement for my requirement?
    Looks like, functional guy is also NOT much idea, here.
    Thanx
    Edited by: Srinivas on Jun 1, 2008 12:15 AM

    no reply

  • Since getting iOS 7 and updates peripheral USB and sync for Itune's is not working.

    Since getting iOS lion and updates I can no longer sync my ipad, or iphone with out being told is no longer authorized on my computer.  Also I can only use the USB port on the back of my Imac for my ipad, and iphone.  Is there a solution?

    I can help with the "IR Receiver" part of your question. That's the piece of hardware inside older MacBook Pros (MBPs) that lets them work with the Apple Remote. "IR" is the standard abbreviation for "infrared," the portion of the electromagnetic spectrum used by remotes for many common devices (stereos, cable boxes, Apple TVs, etc.). The remote sends a coded stream of IR pulses; a corresponding receiver in the target device "sees" the infrared light and decodes it. The MBP's internal IR Receiver sends the received signals to the processor via USB through an internal hub.
    The most recent MacBook Pros have dropped the IR receiver and no longer work with the Apple Remote.

  • The best practice for data mart to different BW System.

    Hi All,
    Would you like to suggest me what i have to do for this case ??
    I have 2 SAP BW systems e.g. BW A & BW B.
    I wanna transfer data from info cube within BW A into info cube within BW B.
    The things that I did :
    1. 'Generate Export Data Sources' for info cube BW A.
    2. Replicate source system in BW B. In BW B, it will show datasource from info cube BW A.
    3. In SAP BW B, I create info package, then I can fetch data from SAP BW A.
    What I wanna ask are:
    1. Could I make it automatically?? Because everytime I wanna fetch data from Info Cube SAP BW A, I must run info package in SAP BW B / what's the best practice.
    2. Could RDA make it automatically ?? Automatic in my case is everytime I have new/update data in Info cube SAP BW A, I don't have to run info package in SAP BW B.
    SAP BW B will automatically fetch the data from Info Cube A.
    If yes, could you give me step-by-step how to use RDA to solve my case please..
    Really need ur guidances all .
    Thanks,
    Best regards,
    Daniel N.

    Hi Daniel,
    You can create a process chain to load your cube in BW A. SImilarly create a process chain in your BW B system to load its cube.
    Now in your system BW A you create a process chain to load your cube. After that you can run automatically the procewss chain in BW B. You can use the Remote chain option for this.
    This will trigger a chain automatically in the remote system.
    Regards,
    Mansi

Maybe you are looking for