Deriving a date using UCM Global Rules

I am trying to use global rules to derive a date/time. When a user selects from a metadata field "status" the value of "Complete" I want the current date and time written out to an "end_Date" metadata field. I have been playing with derived values in Rules but am having some trouble making this work.
Am I on the right track. Should I be able to accomplish this in Rules?
Thanks for your help,
Data_Bitz

Make sure "Complete" is the EXACT value that you're trying to validate against. Remember, option lists have a display value and an internal value. Also check for whitespace. You can try using wildcards as well:
<$if #active.xStatus like "\*Complete\*"$>
<$dprDefaultValue=dateCurrent()$>
<$endif$>
Good luck.
Edit: Had to escape the asterisks to prevent bolding it
Edited by: stellentpmp on Nov 2, 2010 1:50 PM

Similar Messages

  • Comparison of dates using Oracle Business Rule

    Hi to everyone! I have a business rule that has as input parameter a complex attribute with a date field inside named expiration. I want to compare the expiration field with the current date using a comparison function. I don't know where find the current date. The function Calendar.getInstance() return the current date? Any help we'll' be highly appreciated.

    Calendar.getInstance() will give you the current date, and to find out the days between two dates; use Duration.days between(date1, date2) + 1

  • Repeat Data used in Load Rule

    Hi,
    I'm building an report script and a load rule to extract data from one cube into another. The problem is that I want to use one of the fields in the report extract twice. The cube I'm loading to has more demension than the one I'm extracting from. I can determine the projectID in the cube I'm loading to based on the cost center but cost center is also a dimension. I have tried modifying the report script to list the cost center twice but am unsucessful. If it is only listed once, it seems that I can use it for Cost Center or ProjectID but not both. Any ideas?
    Thanks

    In the load rule, open the data file (File->open data file), then select the column that you want to duplicate. Then select field->create using join. It will show you the one column you had selected. click ok and it will duplicate the column, you can then move it where you want, do replacements on it, join it to other columns, or to text columns, tc

  • Use of global data in transformation with parallel processing

    Hi,
    In an upgrade I have a global variable in a routine in 3.5 in a transfer rule. The global variable keeps count of an ID.
    The global variable is used in every data package,because the corrsponding infopackage is set to PSA only, where help says:
    "If you select this processing option and then request processing is done serially during loading, the global data are maintained as long as the process with which the data were processed remains.
    In 7.0 I also use a global variable to do the same thing, but an Infopackage is no longer available, because we use DTP´s now.
    How can I store my global variable so that all parallel processing of the data packages use the same global variable? or can I set an option so that the DTP does serial processing, similar to the infopackage setting in 3.5?
    As a workaround I increased the package size to a very large figure, but I would be more comfortable with a sounder solution.
    thanks

    Hi Max,
      Try to declare your global variable in start routine. Among the lines:
    $$ begin of global - insert your declaration only below this line  -
    $$ end of global - insert your declaration only before this line   -
    Best Regards.
    Javier Gómez

  • Using functional global variables to transfer data between touch screen device and other targets

    Hi,
    We are currently developing a control system that will operate on a cRIO with a touch screen interface. During development we have been using functional global variables (FGVs) to handle most of the settings/data etc, however we have just realised that we may have an issue updating the data on the touch screen, and vice versa.
    Previously we have implemented similar programs using shared variables, however we were wondering if there is anyway to implement FGVs in a similar manner, whether it'd be a combination of FGVs and shared variables.
    Regards,
    Mike
    Software version: Labview 8.2.1

    See this thread on LAVA about sharing FGV's across projects. THis should be extendable to cross paltforms (I believe).
    You also may want to concider an upgrade to LV 8.5.1. We have found a number of performance related issues with LV 8.2.1 and RT.
    If you decide not to upgrade, just remeber what I wrote (above) when you have reason to question the performance.
    Ben
    Ben Rayner
    I am currently active on.. MainStream Preppers
    Rayner's Ridge is under construction

  • Using UCM data controls in ADF application

    Using *.6 Jdeveloper/UCM/Protal Application/UCM
    Calling all the experts in Webcenter/Spaces/UCM/ADF.
    We have a requirement to design search inteface for Documents stored in UCM for Webcenter Portal and Spaces application.
    Our approach is
    Design simple ADF application (no security/templates)
    Bounded task flows using fragments.
    Framents to use UCM data controls (provided by oracle)
    UCM data control definition is extended to use custom metadata.
    package the application.
    Added as ADF library jar to portal application
    Added as Shared Library to Spaces application.
    Question.
    Is this approach sound?
    Does UCM data controls provide all the functionality, for example if we want to read a option list for custom metadata or need custom metadata in return table?
    Can we add other methods to UCM data control provided by Oracle or should we create different one?
    Is there good documentation for such requirements. We have acces Webcenter developers guide which throws some light on UCM data controls.
    Let me know if more clarity is needed...
    Regards

    Hi.
    One of my last components was a custom "Document Explorer" and I used for first time JCR Data Control (Custom UCM Data Control provided by Oracle).
    Sharing my experience I did something similar to you:
    1) Create a ADF Bounded Task Flow and test it individually.
    2) Add generated ADF JAR Library to WebCenter Spaces shared-lib.
    3) Add your Task Flow to Resource Catalog.
    Challenges that I found using custom Data Control:
    1) When items retrieved are content/files all custom metadata can be retrieved OOTB.
    2) When retrieving folders don't provide custom metadata like files. I used RIDC with JCR Data Control to provide all the information relative to folders. If you don't need more information about folders than the name, URI and icons then is OK :).
    3) Using JCR Data Control registers in DataControls.dcx UCM Connection name used when generate JCR Data Control. After deploy your ADF JAR Library in WebCenter Spaces you'll have problems if you don't have a WebCenter Content Connection created with the same name. (Not talking about export UCM Connection with ADF JAR Library, Data Control uses a Connection name by itself :)).
    I added extra functionallity with RIDC (not extending JCR Data Control).
    I hope this help you.
    Regards.

  • How to enter TEXT data using rules files

    Hi,
    How can i enter text data using rules files?
    Thanks,
    A

    If you are talking about planning members with a text data type then it is not possible to load the actual text string using a load rule,essbase just stores a numeric value and the text string relating to that numeric value is stored in the planning applications relational tables.
    If you want to load text directly then you can load data through the planning layer, more information available at - http://download.oracle.com/docs/cd/E17236_01/epm.1112/hp_admin/ch05s02s01.html
    Cheers
    John
    http://john-goodwin.blogspot.com/

  • Check in Similar not populating custom metadata value set by global rule

    I have created a global rule to set the default value for custom metadata field creation date as current system date.This rule is working fine for normal check in but when i am using check in similar functionality of ucm the creation date value is getting populated from original content metadata value instead of default value set by rule.Is there any way to populate creation date as default value set by rule even in the check in similar function.
    Thanks in advance!!!
    Edited by: 906120 on Jan 5, 2012 4:42 AM

    Check-in similar does not evaluate scripts in the "use defaultvalue" field, as it has to copy metadata values. However, you can add script to the "Is derived field" as well. Just don't use dprDefaultValue but dprDerivedValue=
    This script will be evaluated before check-in operation and any value contained in the metadata field will be overwritten with value defined by script.
    Regards,
    Boris

  • Can routine replace "master data attribute of" update rule for performance?

    Hi all,
    We are working on CRM-BW data modeling, We have to look up agent master data for agent level and position for each transaction data. So now we are using "Master data attribute of" update rule. Can we use routine instead of "Master data Attribute of" ? Will it improve the loading performance? Since we have to load 1 lack transaction records , where as we have 20,000 agent details in agent master data.My understanding is, for each record in data package the system has to go to master data table and bring the agent details & store in cubes. Say one agent created 10 transactions, then this option "master data attribute of" will read the agent master data 10 times even though we are going to pull same details for all 10 transactions from master data. if we use routine, we can pull the agent details& storing in internal table removing all duplicates and in update routine we can read the internal table.
    Will this way improve performance?
    let me know if you need further info?
    Thanks in advance.
    Arun Thangaraj

    Hi,
    your thinking is absolutely right!
    I don't recommend to use the standard attribute derivation since it will perform a SELECT to the database for EACH record.
    Better implement a sorted table in your start routine; fill it with SELECT <fields> FROM <master_data_table> FOR ALL ENTRIES OF datapak WHERE OBJVERS = 'A' etc...
    In your routine perform a READ itab ... BINARY SEARCH.... I believe that you won't be able to go faster...
    hope this helps...
    Olivier.

  • Risk Analysis at user level shows nothing in all 3 views though at role level shows risks of global rule set

    I am configuring ARA 10.1 for a ECC 6.0 plug in development system and facing this issue. Risk Analysis at user level shows no data  in all 3 views though at role level shows risks of global rule set. I am using Global rule set. I generated all risks/functions & using connector group as SAP_ECCS_LG not SAP_R3_LG.I activated common, R/3 & ECCS BC sets. Added integration scenario for AUTH. Run all 4 sync jobs multiple times successfully. My system already has decentralised EAM 10.1 implemented & even used in production as BAU. I have checked at both chrome & IE. The misleading thing is that RFC is also working fine & I can see risks in Risk Analysis at role level & risky roles are even assigned to valid users.GRC is at SP4 & accordingly is the ECC 6.0 plug in. Thanks in Advance. Please  consider it urgent.

    Hi,
    Assign ECC connector to SAP_ECCS_LG group.
    Run the programs GRAC_PFCG_AUTHORIZATION_SYNCand GRAC_REPOSITORY_OBJECT_SYNC) in full synch mode(this might take time so better do this in background). Better do it sequentially.Check the logs of the jobs in SLG1 just to ensure everythings fine.
    Run ARA for a specific user and mention the connector for faster output. Ensure this user has the role with risks.Also as explained earlier check the GUID against user id in table GRACUSERROLE and using GRACROLE you can find out the technical name of the role updated in the table. This should be same as the backend role.
    Then run ARA and while doing so please ensure the selection screen doesnt have any unwanted default inputs. If followed correctly , this should be of help.  I am assuming the role analysis yielded correct risks as configured since this would mean that connector have correct actions and basic config is in place.
    Regards,
    Vivek

  • Using labview global in test stand sequence steps

    If I have to pass data between different steps ( only labview vis) within same Main sequence, is it right using LabVIEW globals, rather than TestStand globals?
    in one step, I will be writing data to the Global.
    In a later step, I will be reading the global to access that data.
    Example sequence:
    I am more comfortable with using Labview globals than Teststand variables. Labview is the only adapter in my sequence. 

    If it is just within one sequence, then I would use the TestStand variables.  The reason is because you are limiting your possible reuse of those VIs by using the global.  If you pass in the value, you have a better chance of being able to reuse the VI.
    There are only two ways to tell somebody thanks: Kudos and Marked Solutions
    Unofficial Forum Rules and Guidelines

  • CIO-API: Configuration beeing used when updating rules

    Hi,
    We're currently facing some issues when using the CIO-API to open an exisiting configuration, modify it and then save it again:
    Since we had some minor issues in one of our configurator-extension-rules I changed that rule, updated it in configurator developer, tested it and published the new model to our testing environment. When opening "open" configurations from within order management, my new rules apply. However when using the CIO-API to access an exisiting model, the API still uses the old Model and the old extension-rules.
    My API-Access is as follows:
    CIO cio = new CIO();
    ConfigParameters cp = new ConfigParameters(configHeaderId, configRevNumber);
    Configuration config = cio.startConfiguration(cp, pCpContext);
    //to some modifications
    config.save();
    cio.closeConfiguration(config);
    When I modify this by adding "cp.setRestoreModelId( 281280);" before starting Configuration, my new model is used. However I don't understand why this is necessary and why this is contrary to what happens in order management.
    Could someone clarify how exactly this is working? And: What happens when saving my model using the API? Does it always stay on the "old" level or does it get migrated when explicitly setting the latest restoremodelid on open or are there other circumstances when it's getting 'migrated'?
    And is there an easy way to always use the latest modelId? For now I'd do the following select-statement:
    select newpub.modelid from cz_model_publications oldpub, cz_model_publications newpub,cz_config_hdrs czh
    where czh.config_hdr_id=123708240 and czh.config_rev_nbr=1
    and oldpub.model_id=czh.component_id
    and newpub.top_item_id=oldpub.top_item_id
    and newpub.deleted_flag=0
    But this seems a bit ugly?
    We have no effective-dates or usages on our model if this is of any interest.
    Regards,
    Johannes

    trol how it is restored/relaunched
    1) config_model_lookup_date:-      Date to look up the publication for the configuration Model
    For restored configuration: the saved value of effective_date(cz_config_hdrs), or SYSDATE, as determined by RestoredConfigDefaultModelLookupDate in CZ_DB_SETTINGS
    Don't be surprised if you don't find RestoredConfigDefaultModelLookupDate in the table
    2) config_effective_date:- The date used to filter effective nodes and rules
    For restored(your case) effectie_date(cz_config_hdrs) that was saved is used as default. We can change this using settings
    Now when we launch from OM the above are derived and sent as input parameters(based on your settings). If you want how your isntance is deriving let me know
    For CIO unless you pass default are used.
    Infact when in my code I used the below always to set it explicitly.
    ConfigParameters cp = new ConfigParameters(latestModelId);
    java.util.Calendar modelLookupDate = Calendar.getInstance();
    cp.setModelLookupDate(modelLookupDate);
    --you can add setEffectiveDate to be safe
    Since you explicity set the restoremodelid that is taking care of it. But genereally the same can be achived by setting date parameters.          
    If you have any questions around this area you can reach me directly.

  • Network based and a-gps, data use tests

    i was trying to find this kind of info, but only found others asking.
    i hope this is useful info for others, especially while roaming.
    skip to the bottom for conclusion,
    this is the process i took:
    i called nokia tech support,
    their phone support his horrible,
    the people on the phone don't actually know anything, and have to look up everything you ask them.
    they were the worst.
    i emailed nokia tech support.
    my message was:
    "i would like to know the difference between assisted gps positioning, and network based positioning. these are 2 of the 4 options on my nokia e66 for positioning methods. i would like to know if network based positioning uses internet data from my service provider to obtain my gps position. i am especially concerned with this function while i'm roaming internationally, as i do not want to be charged for roaming data use."
    in about half hour they responded w/ an answer, and then 45 min after that someone else responded w/ an answer.
    the first response, minus the hello's etc. was:
    "In response to your email, Assisted GPS and Network Based positioning methods generate costs related to data traffic so you will be charged for data use. Only "normal" GPS and Bluetooth GPS do not generate additional costs related to data traffic.
    Normal GPS is quite slow and does not work indoors. Therefore, it is recommended to enable at least A-GPS to get a faster position outdoors.
    Network based and Wi-Fi positioning also allow you to update your position inside buildings, but they cannot be used for navigation."
    the 2nd response was:
    "In response to your inquiry, the difference between Assisted GPS positioning (A-GPS) and Network based positioning is that,  A-GPS uses satellites while Network based positioning is based on information of your cellular network environment. It allows you to update your position inside buildings, but they cannot be used for navigation. Both positioning methods generate costs related to data traffic (unless you have configured the device to use Wi-Fi connection to make the internet connection). The cost may vary while you are on roaming. Kindly contact the service provider for information about data transmission costs."
    i can also respond w/ more questions, but instead did some tests, starting with sitting home, indoors, and eventualy moving to a window w/ lots of sky.
    test 1:
    i enabled:
    integrated gps
    network based
    assisted gps
    opened maps,
    set maps to be offline,
    set default internet access point to my home wlan network,
    closed maps,
    cleared all counters and the log on the phone's communication log.
    opened maps
    there is a pink circle around the area i may be in,
    the icon that would show gps satellite strength, changes to show the cell phone antenna icon.
    but the packet data counter is sending/receiving a total of around 3-8kb immediately, and over the course of time, it keeps adding up, every 1-2 minutes, and it is shown as packet data to/from the access point "at&t internet"
    over the course of ten minutes, it's been about 43kb.
    and i can tell on the phone's home screen that the packet data connection becomes active when it does.
    the maps kb indicator still says 0.0kb, with a line through the double arrow packet data symbol.
    test 2:
    now i exited maps,
    cleared the counters and logs
    i turned off a-gps.
    enabled only integrated gps and network based.
    i open maps,
    the same thing is happening as with the assisted gps on, except i connected to my home wlan once, in the middle of connecting to at&t internet 10 or so times.
    everything else is happening the same.
    test 3:
    i exited maps,
    cleared the counters and logs.
    enabled only integrated gps.
    i open maps, and there is no evidence of packet data connection, or wlan connection.
    nothing is showing up in the logs or packet counters, and there is no pink circle around where i may be.
    if i set maps to go online, w/ my wlan as the access point, or at&t internet, their respective icons show up on maps, as well as the phone's home screen.
    the phone's log show respective connections are made but w/ 0.0kb, and the phone's packet counter, as well as the maps kb usage, say 0.0kb.
    test 4:
    set maps to offline, w/ my wlan as access point.
    exited maps,
    turned phone off for a bit to test gps from cold start.
    turned on phone,
    cleared counters/logs,
    enabled only assisted gps and integrated gps.
    open maps, there's no pink circle around where i may be, just the red dot of where it thought i last was.
    there's about 4-5kb of data transfer,
    then i bring the phone to the window where there's a lot of clear sky, and my location is found pretty fast.
    i know that w/o assisted gps, w/ only integrated gps, it takes a while to get the gps signals.
    i went back and forth to/from the window, and there was packet data use again only once more, not continuously like w/ network based positioning.
    so to conclude my tests,
    setting the maps to be offline, still uses a-gps and network based positioning.
    network based uses a more data, continuously,
    which is the opposite of what i thought, i thought a-gps would use more, so i was using network based on and a-gps off while in canada. whoops.
    using assisted gps uses about 5-10kb, which, when roaming, according to at&t, they charge 1.95cents/kb.
    might be worthwhile for getting a quick gps signal, nothing worse than waiting forever for it to connect.
    the only way to be completely clear of data use is disable network based and assisted gps, or set the phone to offline mode.

    A-GPS: This uses information from cell network to get a rough idea of where you are. It then uses this information to figure out which satellites to look for. Result is a faster satellite lock. So A-GPS uses a few kB of data to get an initial position, but navigation is done using GPS satellites, which does not require ANY data connection.
    Network based: This uses information from the cell network to plot the current position when GPS satellite signal is not available, ie indoors, in a tunnel, etc. Position info derived from the cell network is not as accurate as that from GPS satellites, but serves as a stopgap until GPS signal can be re-acquired. Network based positioning used your data connection EACH TIME the satellite lock is lost. So the amount of data used is dependent on the quality of access you have to the GPS satellites. If you don't lose the connection to the GPS satellites, then you won't use any data.
    The ‘offline’ option within the Nokia Maps app only refers to street, POI, etc searches, NOT to A-GPS, Network based positioning or to connecting with the GPS satellites.
    If you want to navigate without using ANY data connection, set A-GPS off AND set Network based positioning off AND set the Internet option in Nokia Maps to Offline. Do this and you will not use any of your data connection allowance when navigating. There is no need to put the phone into flight mode.
    If I've helped you, you can thank me by clicking the green 'kudos' star on my post. Cheers.

  • Adding data to Manual Distribution Rule from DIAPI

    Hi
    I am searching for the object to insert the data to OMDR (manual distribution rule) using DI API. Can any one help me out.
    I found many questions related to adding data to manual distribution rule using DIAPI which were not answered.
    Thanks in Advance.

    hi.
    Just check the below posts.
    object type of distribution rules
    Distribution rules setup..

  • Use of global temporary tables in Procedures

    Hi
    I am using global temporary tables in the procedures. Loading data in the same table through many procedures. I am fetching the data from the global temporary table in PRO-C by a cursor. Will this degrade performance?
    Please help me..
    Thanks in Advance...

    Will this degrade performance?That depends... in comparison to what?
    Loading data into temporary tables will generally be more efficient than loading data into permanent tables because Oracle needs to do less to protect this data since it is inherently transient. On the other hand, loading the data into a table in the first place tends to be more expensive than alternatives like using a single SQL statement, a pipelined table function, or an in memory collection.
    Justin

Maybe you are looking for

  • The cryptic error - element cannot be found in a collection ?

    I have encountered this error in the script task a few times. The causes are - Variable is present in code,  but not added to task list OR, you try to create a DataColumn with the wrong column name, OR the variable in the task list has been deleted.

  • Forms Error Message

    Hi, I am using Forms 9i Version 9.0.2.9.0 and Database Oracle 9i Enterprise Edition Release 9.0.1.1.1. Database and Developer are in the same computer (Windows XP Professional). Folders are ora90 and 9iDSHome respectively. OC4J is running properly. M

  • Missing pages when printing in Adobe Reader 9?

    I'm trying to print a chapter in a PDF document using the Bookmarks Navigational Panel but whenever I print the chapter (i.e. right-click on the chapter and select Print Pages), some pages are skipped and not printed. I have tried to re-install Adobe

  • HT4436 Changing my icloud email address

    I am getting married soon and want to change my iCloud email address. How can a person do this?

  • Webservice to JDBC to SAP System to webservice using BPM

    Hi Frnz My requirement is i need to get data from webservice....... and i need to push that data to some Database... and the response from the Database to SAP system......and from that SAP system  the respone should come to the Webservie which is the