DPS6: hiding attributes for entries not matching a specific objectclass

I'd like to be able to hide some attributes, but only for entries not matching a given objectclass.
DPS 6 allows data hiding rules to apply to a specific subset of entries, but the subset can not
be defined thanks to a negative filter such like (!(objectclass=MyAuxiliaryObjectClass)).
If I hide the attributes at the view level, then it takes precedence over any "allow rules", so I don't see how
I can achieve this.
Any idea ? support plan ?

Just to add some more details, I've tried to use a virtual ACI in order to define my filter but there's no "targetfilter" support in DPS, and with different other ACI combinations, it can not work because there's no implicit deny ACI in DPS:
Even with only "allow dpsaci's" , attributes not specified in those dpsaci's are still allowed by the viewable-attr parameter of the view. Moreover, if I use the viewable-attr/non-viewable-attr view parameters, then the not explicitly allowed/explicitly denied attributes defined at the view level take precedence over the dpsaci's .

Similar Messages

  • Rectify the error mesage "Account determination for entry *** not possible"

    When I run the program ZMR11-Maintain GR/IR Clearing account, I get the below said error message:
    +Account determination for entry **** not possible+
    How do I rectify them?
    Please advise.

    Check the long text of the error message. It should ideally give details of missing entry in account determination.

  • Error msg in MIGO: Account determination for entry not possible

    hello when i am doing a goods movement, i get the message Account determination for entry DART WRX DFGX ___ 3200 not possible.

    Dear,
    check one moresettings VC/Plant assigned to Valuation grouping code through T.code OMWD.
    Regards,
    Kishore K

  • Accounting doc entries not matching with the the entries from j_1ipart2 entries

    Hi,
    After creating PO, I have done MIGO and J1IEX. Now, posted MIRO. The end screen shows a number stating accounting document created.
    I checked the accounting doc in fb03 and the document is displayed over there. I also checked the table for values captured in PART2 of excise. The value displayed in table is not matching with the value displayed in accounting document. Can anyone help?

    Hello Tejas,
    Here is the updated version of the situation.
    After the transaction J1IEX, I got a message saying an 'Accounting Document' has been created. But the accounting document number does not match with the original values. I checked all the tables and found out that the accounting document is not created.
    In this case, I thought of creating an accounting document manually and adding them directly to the table.
    Here are my questions,
    -Can I manually create an accounting document for multiple line items?
    -Few tax values should be split and should go to 2 different GL accounts. so how to do this if I am creating it manually
    -Or can I delete the values captured during 'J1IEX' in the table and reopen the GR so that the t-code j1iex can be processed again?

  • Search results for podcasts not matching iTunes profiles

    There are podcasts on iTunes that have different titles and artwork in search results than their iTunes profile pages. For example, the podcast listing for LightTalk podcast: http://itunes.apple.com/WebObjects/MZStore.woa/wa/viewPodcast?id=333378833 displays the correct title and artwork, but when searching the iTunes Store for 'LightTalk', the artwork is missing and the title is incorrect (looks like the title is the author's name in the search results). This podcast was submitted to iTunes about 2 weeks ago. There are other podcasts having the same problem and suspect the problem is isolated to only new listings.
    This is a problem on both Windows and Mac versions of iTunes.

    I to have the very same problem! search results do not match the feed. Its been 1 Week now, i've added 2 more podcast since feed was submitted, the feed page works perfect, picture, artist, desciption, and updates within hours... perfect!
    The problem is how often dose the search index get updated?
    Regards

  • Which operator  should i used for loading not matched records

    HI ALL,
    I am just evluate OWB tools.
    and i met a question.
    I need to join two tables A and B with Condition A.id=B.id
    For example,
    A has 1000 records, and B has 100 records,
    what i need to do is
    (1) to get matched 100 record from B to a new table C,
    (2)to get the not matched 900 records from A to another table D.
    By the requirement like above, which operator could i use for?
    I tried to use joined but can not get the record of (2).
    Thanks for your attention.
    Regards

    Hi,Sutirtha
    Thanks for your attention.
    because joiner has only one output group.
    so what you mean is use joiner operator twice?
    One joiner userd to get the matched records,
    and the other used to get unmatched records?
    Thanks
    Rachel

  • No ACL deny logs for Traffic not matched by Static Object NATs and ACL. Need Help.

    I start noticing that I do not see any denied traffic coming in on my ACL.  To better explain, lets say I have this config.
    ### Sample Config ###
    object network webserver
    host 192.168.1.50
    nat (dmz, outside) static X.X.X.X service tcp www www
    access-list inbound extended permit ip any4 object webserver eq www
    If I generate a traffic from the outside let's say a traffic that is trying to access X.X.X.X via TCP Port 8080 which obviously does not have any NAT entry to it going to my DMZ, I don't see the ACL denies it anymore but instead comes back with a Drop Reason: (nat-no-xlate-to-pat-pool) . On the packet trace I got this. (Below) it seems that does not even hit the ACL as there is no xlate found for it, at least to what the drop reason says.
    Phase: 1
    Type: CAPTURE
    Subtype:
    Result: ALLOW
    Config:
    Additional Information:
    MAC Access list
    Phase: 2
    Type: ACCESS-LIST
    Subtype:
    Result: ALLOW
    Config:
    Implicit Rule
    Additional Information:
    MAC Access list
    Phase: 3
    Type: ROUTE-LOOKUP
    Subtype: input
    Result: ALLOW
    Config:
    Additional Information:
    in   0.0.0.0         0.0.0.0         Outside
    Result:
    input-interface: Outside
    input-status: up
    input-line-status: up
    output-interface: Outside
    output-status: up
    output-line-status: up
    Action: drop
    Drop-reason: (nat-no-xlate-to-pat-pool) Connection to PAT address without pre-existing xlate
    Before, using a regular Static PAT on ASA Versions 8.2(5) below, I could get the deny logs (ASA-4-106023). Generally, I use these logs, and are quite important for us specially during auditing.
    My question is how can I generate logs for these type of dropped traffic on the ASA 9.1 Version? 
    Any comments/suggestions are gladly appreciated :)
    Regards,
    John

    I believe, but am not 100% sure, that the reason you are not seeing the ACL drop but a no NAT matched is because of the changes from 8.2 to 8.3 in the order of how things are done.  In 8.3 and later you need to secify the real IP address when allowing packets in, and this is because NAT happens before the ACL is matched.  So since there is no match on the NAT the packet is dropped then and there, never reaching the stage where ACLs are checked.
    As to seeing drops in the ACL log...You might want to try adding an ACL that matches the NATed IP...but I don't think you will have much success with that either.  My guess is that there is no way around this...at least no way I know of.
    Please remember to select a correct answer and rate helpful posts

  • Current Period" attribute for Variants not working correctly

    Hi Expert,
    Pls i need your advice on the issue screenshot below.
    When creating variants, a dynamic screen selection can be chosen for “Current Period” that will pre-populate begin and end dates. The problem is that, this attribute gives calendar month dates, not our company accounting period dates i.e. 4-4-5.
    Can this be corrected?
    I have an option to create the fiscal year variant to fix the problem but am not sure on how to go about that.
    Pls advice.
    Regards
    Femmy John.

    Hi Marssel,
    Thank you for the quick respond.
    Kindly explain to me the importance of using this field period and execute same setup?
    will this populate our company accounting period dates i.e. 4-4-5.
    I just want to get the fact right before going to the customer.
    Regards
    Femmy John

  • Retain Pattern Match Attribute for Next Pattern Match Calculation

    When a MATCH_RECOGNIZE PATTERN match occurs, is there any way to retain the underlying event for the next pattern match?
    It's like I want to replay the underlying event (that triggered the match) into the next "batch" of events.

    Here is an approach
    1) Use a (local) cache to hold the "lastMatchedTime"
    2) Augment each input event with the "lastMatchedTime" by first (left outer join) joining with cache
    3) Run the MATCH_RECOGNIZE on the augmented stream
    4) Use a match to update the cache
    Here are the queries that I tried -
      <view id="S">
            <![CDATA[
                 RSTREAM (
                    SELECT
                           id,
                           obs,
                           epoch,
                           nvl(C.lastMatchedTime, 0) as lastMatchedTime
                    FROM
                           InputChannel[now] AS I LEFT OUTER JOIN lastMatchCache AS C
                    ON
                           C.keyValue = "key"      
            ]]>
          </view>
          <view id="v1">
            <![CDATA[                 
                      SELECT
                              T.id,
                              T.obsTotal,
                              T.elapsedTime,
                              T.firstts,
                              T.lastts
                      FROM S MATCH_RECOGNIZE (
                              PARTITION BY id
                              MEASURES
                                        id as id,
                                        sum(obs) as obsTotal,
                                        B.epoch - decode(B.lastMatchedTime, 0L, min(epoch), B.lastMatchedTime) as elapsedTime,
                                        decode(B.lastMatchedTime, 0L, min(epoch), B.lastMatchedTime) as firstts,
                                        B.epoch as lastts
                              PATTERN ( A+? B )
                              DEFINE
                                       B AS (epoch- decode(lastMatchedTime, 0L, min(epoch), lastMatchedTime)) > 20*60*1000
                      ) as T
            ]]>
          </view>
          <query id="qcache">
             <![CDATA[
                 SELECT
                       "key" as keyValue,
                       lastts as lastMatchedTime
                 FROM
                       v1
             ]]> 
          </query>
           <query id="q1">
             <![CDATA[
                 SELECT
                 FROM
                        v1
             ]]> 
          </query>
        Also, include in the config file
    <channel>
           <name>CacheChannel</name>
           <selector>qcache</selector>
       </channel> 
       <channel>
           <name>OutputChannel</name>
           <selector>q1</selector>
       </channel>The EPN that I used was
      <wlevs:event-type-repository>       
            <wlevs:event-type type-name="CacheEvent">
                <wlevs:class>com.bea.wlevs.event.example.helloworld.CacheEvent</wlevs:class>
            </wlevs:event-type>
            <wlevs:event-type type-name="InputEvent">
               <wlevs:properties>
                  <wlevs:property name="id" type="int"/>
                  <wlevs:property name="obs" type="int"/>
                  <wlevs:property name="epoch" type="bigint"/>
               </wlevs:properties>
            </wlevs:event-type>
            <wlevs:event-type type-name="OutputEvent">
               <wlevs:properties>
                  <wlevs:property name="id" type="int"/>
                  <wlevs:property name="obsTotal" type="int"/>
                  <wlevs:property name="elapsedTime" type="bigint"/>
                  <wlevs:property name="firstts" type="bigint"/>
                  <wlevs:property name="lastts" type="bigint"/>
               </wlevs:properties>
            </wlevs:event-type>
        </wlevs:event-type-repository>
        <wlevs:adapter id="InputAdapter" provider="csvgen">
             <wlevs:instance-property name="port" value="9061" />
             <wlevs:instance-property name="eventTypeName" value="InputEvent" />
             <wlevs:instance-property name="eventPropertyNames" value="id,obs,epoch" />
        </wlevs:adapter>     
        <wlevs:channel id="InputChannel" event-type="InputEvent" >
            <wlevs:listener ref="helloworldProcessor"/>
            <wlevs:source ref="InputAdapter"/>
        </wlevs:channel>
        <wlevs:caching-system id="local-caching"/>
        <wlevs:cache id="lastMatchCache" name="lastMatchCache" key-properties="keyValue" value-type="CacheEvent">
           <wlevs:caching-system ref="local-caching"/>
        </wlevs:cache>
        <!-- The default processor for OCEP 11.0.0.0 is CQL -->
        <wlevs:processor id="helloworldProcessor">
           <wlevs:listener ref="CacheChannel"/>
           <wlevs:cache-source ref="lastMatchCache" />      
        </wlevs:processor>
        <wlevs:channel id="CacheChannel" event-type="CacheEvent">
           <wlevs:listener ref="lastMatchCache"/>
        </wlevs:channel>
        <wlevs:channel id="OutputChannel" event-type="OutputEvent" advertise="true">
            <wlevs:listener>
                <bean class="com.bea.wlevs.example.helloworld.HelloWorldBean"/>
            </wlevs:listener>
            <wlevs:source ref="helloworldProcessor"/>
        </wlevs:channel>And the output that I get seems to match your requirement -
    eventType=OutputEvent object=q1  v1.id=4143, v1.obsTotal=9, v1.elapsedTime=7228000, v1.firstts=1199311210000, v1.lastts=1199318438000
    eventType=OutputEvent object=q1  v1.id=4143, v1.obsTotal=17, v1.elapsedTime=30935000, v1.firstts=1199318438000, v1.lastts=1199349373000
    eventType=OutputEvent object=q1  v1.id=4143, v1.obsTotal=7, v1.elapsedTime=41436000, v1.firstts=1199349373000, v1.lastts=1199390809000
    eventType=OutputEvent object=q1  v1.id=4143, v1.obsTotal=13, v1.elapsedTime=5904000, v1.firstts=1199390809000, v1.lastts=1199396713000Edited by: Anand Srinivasan on Sep 30, 2010 12:49 AM
    Edited by: Anand Srinivasan on Sep 30, 2010 12:50 AM

  • Calender Entries not matching on the iCloud of iPad2 and Windows PC 7

    I have iCloud on my iPad2 and Windows 7 Laptop, the calender entries are reflected with an hour's time difference.
    I have checked the Timezones on both the devices and they have the same Time zone.
    Don't know why the time lag is present. Kindly help.

    Have you checked the timezone settings in iCloud?
    Login https://www.icloud.com/ and press your name in the upper right corner. There you find TimeZone setting and there's might be your problem

  • While doing PGI get error account determination for entry

    while doing PGI get error account determination for entry    not possible message no M8147

    Hello,
    Seems like you are missing entry in Trx OBYC - GBB/VAX or GBB/VAY. The long text of the error message would tell you where to maintain the entries. If that is VAX or VAY you would need to maintain the 'Cost of Goods Sold a/c there according to the valuation class.
    Kind Regards // Shaubhik

  • The SYSVAL table entry for the database version (16) does not match the required version

    We upgraded from Tidal Enterprise Scheduler (TES) 5.31 to 6.1 on fresh Windows x64 2008 R2 servers.  I did a fresh install but our DBA restored a copy of our database from pre-prod which was at 5.31.  I get the error below.  Note, SQL went from 2005 to 2012 during this upgrade.  Do I need to change a value in this table to reflect the SQL change?
    [04/29 12:47:14:198]:TIDAL Enterprise Scheduler: version 6.1.0.133
    [04/29 12:47:14:198]:Java version: 1.8.0
    [04/29 12:47:14:198]:Java Virtual Machine version: 25.0-b70
    [04/29 12:47:14:198]:Start Time : 04/29/14 12:47:14:198
    [04/29 12:47:14:198]:----------------------------------------------------------------------------
    [04/29 12:47:14:198]:Database URL :jdbc:sqlserver://SQL2012-Host:1433;responseBuffering=adaptive
    [04/29 12:47:14:198]:Database Driver :com.microsoft.sqlserver.jdbc.SQLServerDriver
    [04/29 12:47:14:198]:Maximum number of log files = 100
    [04/29 12:47:14:198]:Added a LogFile called 'RegularFile'
    [04/29 12:47:14:198]:LogManager: setting default log
    [04/29 12:47:14:214]:Retrieved a LogFile called 'RegularFile'
    [04/29 12:47:14:495]:MessageBroker: Instantiated TcpTransportServer (URI = tcp://0.0.0.0:6215)
    [04/29 12:47:16:975]:Retrieved a LogFile called 'RegularFile'
    [04/29 12:47:17:272]:The SYSVAL table entry for the database version (16) does not match the required version (23). Shutting down.
    [04/29 12:47:20:282]:
    [04/29 12:47:20:282]:
    [04/29 12:47:20:282]:Shutting down the application

    I had this error last night while applying a patch to 6.0. However in my case the version was 21 versus a required version of 23.
    As far as i know 6.1 is not compatible with the 5 db schema, so you will need an update / migration plan for the db also.

  • Data type of the base attribute or the base value does not match...

    ...the assigned expression.
    Hello all,
    I always get the Error
    +<ERROR+
    TEXT="'DWH.CUB_REGISTRATIONS_AW.REGISTRATIONS': XOQ-02517: Der Datentyp des Basisattributs oder der Basisgröße stimmt nicht mit dem zugeordneten Ausdruck überein.
    XOQ-01400: Ungültige Metadatenobjekte"/>
    The English message must be something like this:
    The data type of the base attribute or the base value does not match the assigned expression.
    when I run my mapping. The attribute REGISTRATIONS is NUMERIC (12,2) in the Cube and I map a NUMERIC(12,2) constant in it.
    I use a simple OWB-Mapping for loading, but I don't understand why it doesn't function. Other mappings where the attributes are out of a
    table I put in a cube are running well.
    I tried different things, but nothing fixed my problem. Any idea ?
    Thanks a lot for help
    Michael

    Technically this is a 'warning' from the server, not an 'error'. This means that the change you made should have been submitted, but you get an warning message on the client. AWM would suppress this warning, but evidently OWB does not. Can you switch to use AWM?
    Here is the definition of the warning along with 'cause' and 'action' sections. (Unfortunately these sections are not translated into German for some reason.)
    >
    02517, 0, "The data type \"%(1)s\" of the base attribute or base measure is different from the mapped expression \"%(2)s\"."
    // *Cause: Either the base attribute or base measure with the mapped expression was set to an inconsistent data type, or it was mapped to an expression of a different data type from its fixed data type.
    // *Action: When changing a mapped expression for a base attribute or base measure, ensure that the expression has the same data type; otherwise, set the data type of the base attribute or base measure to NULL first. When a base attribute or a base measure has an existing mapped expression, do not set it to a different data type.
    >
    It is probably safe to ignore this warning, but if you can post the relevant XML for the cube, then will probably be able to spot the problem. I assume that REGISTRATIONS is a measure in the cube CUB_REGISTRATIONS_AW, so this is what you can look for in the XML:
    (1) The definition of the base measure along with the datatype. It should be something like this
    <Measure>
      <BaseMeasure
        SQLDataType="NUMBER(12,2)"
        ETMeasureColumnName="REGISTRATIONS"
        Name="REGISTRATIONS">(2) The mapping info for the measure, which should looks something like this:
    <MeasureMap
      Name="REGISTRATIONS"
      Expression="...">
      <Measure Name="REGISTRATIONS"/>
    </MeasureMap>I don't know if you can get the XML directly from OWB. If not, then DBMS_CUBE.EXPORT_XML should work (assuming you are in 11.2). You could also attach AWM and save the cube to an XML template.

  • General ledger does not match profit and loss for current year

    Closed 2008 and it created some unexpected journal entries. These JEs show up as beginning balances in the 2009 GL. Also net income does not match profit for current period on balance sheet. Tried reversing the unexpected journal entries and this caused the 2009 retained earnings to be more out of balance. How do I fix this?

    As far i understand you have done Balance Carry Forward by FAGLGVTR.
    and your opening balance in next year is not correct as desire in Retain earning account.
    If the case so then kindly check that you have transfer amount from retain earning account to and reserve account in previous year.
    Regards
    Shayam

  • Attribute member  datatype is not matching in Essbase Studio

    Hi All,
    When I was trying to build the ASO cube by adding attributes it throwing below error
    Failed to deploy Essbase cube.
    Caused by: Cannot end incremental build. Essbase Error(1007083): Dimension build failed. Error code [-1]. Check the server log file and the dimension build error file for possible additional info.
    \\Outline verification errors:
    \\Attribute member datatype is not matching with Attribute dimensions datatype.
    \\Illegal datatype for attribute member
    \\Attribute member (Name) datatype is not matching with Attribute dimensions datatype.
    \\Illegal datatype for attribute member Name
    Here every thing is "String only" and i verified in backend it will show ing datatype as "Varchar".
    Can someone assit me how to resolve this error.Thanks in Advance
    Regards,
    SatyaB

    Did you tried CASTing your source data to match it attribute dimension type? Just CAST all source data before its loaded and it should help resolve any data mismatch type errors.
    I think I have an integration where we use VARCHAR2 type data mapped to string dimension.

Maybe you are looking for