Date stamping Cubes

<p>Our firm has a requirement that the date the cube was lastupdated be included in any queries made of the cube. Does anyonehave an idea how this might be handled?</p><p>To be useful the date should natively show as a date without theuser having to "drilldown" on a field.</p><p> </p><p>Thanks</p>

One way I've overcome this is to have a dimension called Report_date. Every load I change the alias to be today's date. Then when the user retrieves, they use aliases and have report date in the report. It adds no overhead to the cube since it is a single member dimension<BR>

Similar Messages

  • Data in cube is different from psa in the production system

    hi friends
    this is very urjent ,the data is fine and  same as r/3 in psa.for example i have sales for one article ie billing
    2lis_13_vditm . which picked the data from r/3 when i see the records in the psa they r good when tried to see the same record in cube record is not avalible in the cube . only few records are filtered out inbetween psa to cube which is leading to lot of data inconsistency . they r no routines which can filter out the data . only standard sap routines .which updates the data to cube . what could be the problem . any help is apperciated and helpful and will be rewarded. thanks in advance for kind replays.

    veda,
    In a cube the data gets added up for similar records -
    Do you have the same number of records in PSA and Cube ?
    if yes - then maybe there exist similar records and the KF is getting summed up in the result.
    Also how did you search for the same record in the cube ? since the characteristics go into the dim tables and all that is in the fact table is only dim ids and KFs ...
    Do one thing - can you try a report on the cube and check if the data is getting summed up.
    Or another workaround - put the PSA Data into an Excel file ( flat file ) and then upload it into an ODS with the same records in Production - you will know if multiple records exits and thereby find out what the problem is due to. ( Bad workaround - but cannot do it in Production)
    Arun

  • Data mart cube to cube copy records are not matching in target cube

    hI EXPERTS,
    Need help on the below questions for DATA Mart-Cube to cube copy(8M*)
    Its BW 3.5 system
    We have two financial cube.
    Cube A1 - Sourced through R/3 system (delta update) and Cube B1- Sourced through A1 cube.(Full update). These two cubes are connected through update rules with one to one mapping without any routines.Basis did a copy of back-end R/3 system from Production to Quality server.This happened approximately 2 months back.
    The Cube A1 which extracts delta load from R/3 is loading fine. but for the second cube, (extraction from previous cube A1) i am not getting full volume of data instead i m getting meagre value but the loading shows successful status in the monitor.
    We  tried through giving conditions in my infopackage (as it was given in previous year's loading) but then also its fetching the same meagre volume of data.
    To ensure that is it happening for the particular cube, we tried out in other cube which are sourced thro myself system and that is also getting meagre data rather than full data..
    For Example: For an employee if the data available is 1000, the system is extracting randomly some 200 records.
    Any quick reply will be more helpful. Thanks

    Hi Venkat,
                  Did you do any selective delitions in CUBEA1.
    first reconcile data cube1 & cube2 .
    match totals of cube1 with cube2.
    Thanks,
    Vijay.

  • Delete data from Cube which is uploaded to Open hub

    Hello Friends I am required to delete the data from GL Cube, But it is connected to 2 Open Hub
    which has one as table /BIC/OHZZIGL_C10 defined in Open hub destination. I deleted the data from this table in SE14, but still when I try to delete the data from cube it throws the below error
    Please help me how to now delete the data from cube, when should i go in open hub t delete all the uploaded requests.
    Request 1.882 already updated in target 3.913 by DTP request DTPR_4CU6JHRJ7MY889A7XBUJ0GDVK(ZZIGL_C10)
    Thanks
    soniya

    just check on your datamart status for that request. Go to the manage screen of your infocube and check where all that request got updated using datamart status.

  • Date Stamping Photos

    I am newbie to Lightroom but am quickly loving this program.  One of the issues that led me here is date stamping.  Our dSLR does not allow for date stamping (Nikon D40x).  What it does is retains the "capture date" in the metadata ( i beliee most cameras do).  We take approximately 1,000 photos a month and would love to have them date stamped.
    Our processing is done through Kodak, who will not read the capture data from the metadata and stamp on the photo.
    My next option was thir party software - in comes Lightroom 3.  Well reading through the forums I found this was possible via LR Mogrify 2.  It was installed and it works great.  The amount of time I save vs the Gena Photstamper is incredible.  The problem I am having now is that Kodak auto crops the pictures before developing which crops the date stamp (located on the bottome right hand corner).
    I have selected on the Kodak site "print entire picture" which elliminates the cropping but leaves white bars on the ends of the picture.  The reason for this per the Kodak rep is due to photo size.  My next step was to go back into Lightroom and set the export size to 1200 X 1800.  The result was something close but not exactly that dimension.  Which still leaves me with either vertical or horizontal bars on the picture when processed.
    If anyone has encountered anything similar or if you have any possible solutions can you please help.

    May I suggest that you buy a camera that will imprint the date on your
    photo and be done with it. There are many point and shoots with good quality
    that will do this very thing.
    Keep it simple..........
    Rich

  • How to enter a date stamp using formscentral?

    I've been trying to enter a date stamp using formscentral but have no such luck. Whenever I enter one in using acrobat pro I get the error that I don't have sufficent privleges due to adobe keeping the integrety of the form...Anybody have any ideas?

    See this FAQ which explain how to modify a FormsCentral PDF in Acrobat
    http://forums.adobe.com/docs/DOC-3661
    Gen

  • Loading data from Cube to ODS

    Hi All,
    I have a cube which contains projects data. I want to load the data from cube to a write optimized ODS. In the cube for a project i have multiple records. But while loading the data to ODS, I have to summarise the data in such a way that i should have one record for a project i.e if the cube has data as follows:
    Project Cost
    abc      100
    abc       200
    abc       300
    Then in the ODS the record should be
    Project    Cost
    abc        600
    How do i achieve this ?
    Thanks,
    Satya

    Hi Satya,
    Generate export datasource from Cube ( Right click on cube ->generate export datasource).
    Create update rules for the ODS and in the update rule of Cost , you will find two options
    1. overwrite
    2. addition.
    Choose addition and activate update rules. ..This will work exactly same way you wanted.
    One more thing to mention ,Put project info object in to key fields of ODS and rest all info objects in to data fields.
    Hope that helps.
    Regards
    Mr Kapadia
    Assigning points is the way to say thanks in SDN.
    Message was edited by:
            Mr Kapadia

  • Urgent: Error in uploading data in cube after adding new Key figure

    Hi Guys,
    I have added 1 key figure (DEC) to my existing cube. and proceeded as below:
    1. Deleted all data from cube
    2. activated the cube
    3. set Constant 1 for this key figure
    4. activated the transformation
    5. reated a new DTP
    But when i execute it, it gives me error. in updating to infocube ZSDBILL:
    Error while updating to target  ZSDBILL
    Process Terminated.
    all other steps in DTP are succesful like: extraction, error handling and transformation.
    There is no details in long text of these errors.
    Can anyone suggest whats the error stands for or what i am missing.

    Once logout and login back then again activate cube,transformation and run the DTP.
    Hope it helps you

  • Dead lock error while updating data into cube

    We have a scenario of daily truncate and upload of data into cube and volumes arrive @ 2 million per day.We have Parallel process setting (psa and data targets in parallel) in infopackage setting to speed up the data load process.This entire process runs thru process chain.
    We are facing dead lock issue everyday.How to avoid this ?
    In general dead lock occurs because of degenerated indexes if the volumes are very high. so my question is does deletion of Indexes of the cube everyday along with 'deletion of data target content' process help to avoiding dead lock ?
    Also observed is updation of values into one infoobject is taking longer time approx 3 mins for each data packet.That infoobject is placed in dimension and defined it as line item as the volumes are very high for that specific object.
    so this is over all scenario !!
    two things :
    1) will deletion of indexes and recreation help to avoid dead lock ?
    2) any idea why the insertion into the infoobject is taking longer time (there is a direct read on sid table of that object while observed in sql statement).
    Regards.

    hello,
    1) will deletion of indexes and recreation help to avoid dead lock ?
    Ans:
    To avoid this problem, we need to drop the indexes of the cube before uploading the data.and rebuild the indexes...
    Also,
    just find out in SM12 which is the process which is causing lock.... Delete that.
    find out the process in SM66 which is running for a very long time.Stop  this process.
    Check the transaction SM50 for the number of processes available in the system. If they are not adequate, you have to increase them with the help of basis team
    2) any idea why the insertion into the infoobject is taking longer time (there is a direct read on sid table of that object while observed in sql statement).
    Ans:
    Lie item dimension is one of the ways to improve data load as well as query performance by eliminationg the need for dimensin table. So while loading/reading, one less table to deal with..
    Check in the transformation mapping of that chs, it any rouitne/formula  is written.If so, this can lead to more time for processing that IO.
    Storing mass data in InfoCubes at document level is generally not recommended because when data is loaded, a huge SID table is created for the document number line-item dimension.
    check if your IO is similar to doc no...
    Regards,
    Dhanya

  • Getting an error while trying to load data into cube

    Hello,
    I am trying to load data into cube but it gives me following error:
    fiscal year variant not expected.
    Can someone tell what could be done on fiscal year
    Cheers
    Jim.

    Hi,
    May be you have not mapped anything to fiscal variant or not included fiscal variant in the cube but you have used fiscal period inside it.
    You should load fiscal variant if you wanna use fiscal period.
    Map fiscal variant to a constant K4 or from the source if present and then see the result.
    Generally it is K4 except for FI-CO cubes .
    In the transfer rules, select field '0fiscvarnt' 'Tp' click
    Enter K4 for Constant................Save and Activate
    Iam sure you'd fix the issue.
    Regards,
    Ray

  • While deleting the data from cube..getting the following erros...

    while deleting the data from cube the below error occured in develoment..any help
    i have 4 requets.....2 are deltedd but no records for those...
    2 are not deleted and they have some data..
    please give me u r inputts pelase
    Performing check and potential update for status control table
    Message no. RSM1490
    Diagnosis
    If data is loaded into an InfoCube, or existing data is edited (aggregated/compressed/deleted/got from a DataMart), then there is a change in  the potential reportability of the data , or the possibility of deleting data by request, or of aggregating or compressing.
    This status of each one of these Cubes is stored in a status table, that is updated when there is any change to the status of a request in the Cube.
    The system now analyzes the requests in the Cube and compares the calculated status with the status table.
    If deviations from the status table arise then you are given the option of adjusting the status table.
    System Response Procedure Procedure for System Administration

    Hi,
    Did you tried by right click on cube and delete data? are those requests are compressed or roll up?

  • While loading transaction data into cube what are the tables generats

    Hi,
    while loading transaction data into cube what are the tables normally generats.

    Hi,
    Normally datas will be loading to 'F'- Fact tables (/BIC/F****) *** - cube name..
    When you do compress the request the data will be moved into E tables.
    Regards,
    Siva.

  • How to do reconcilization of ODS data and CUBE data

    Hi All,
    How to do reconciliation of ODS data and CUBE data,I know how to do reconciliation of R/3  data and BW data.
    Regards.
    hari

    Hi,
    create a multicube based on your ODS and cube, identify commen characteristics and perform the key figure selections; create then a query showing both information from the cube and ODS with perhaps some formula showing differences between key figures.
    hope this helps...
    Olivier.

  • What is the difference between methods of deleting data from cube

    Hi,
    Is there any differences, pros and cons for deleting data from cube by:
    1) deleting the specific request loaded into the cube
    2) using selective deletion with specific characteristic values
    Assumption for the above is both the options will delete the same data set.
    Thanks and regards.

    Hi BW Beginner,
    Yes, of course, there are differences.
    Selective Deletion:
    PROS: You can use to delete data that satisfies a certain criteria (e.g. some characteristic values).
    CONS: I think this locks the whole cube during deletion
    Deletion by request id:
    PROS: You can easily delete all data that belongs to a certain request id (i.e. you can easily delete a bad load).
    CONS: If you have already compressed the cube, you can not use deletion by request id to delete a certain request.
    Hope this helps.

  • How do I add time/date stamp to my screen capture file names?

    I'm on mac osx.
    I'm trying to add time/date stamp to my screen capture file names. (at the moment it's just 'Picture 1' etc..)
    I've tried the following command  in terminal but have not had success. please help!!
    defaults write com.apple.screencapture name "datestamp" at "timestamp"
    killall SystemUIServer

    Surely someone else will provide a better solution. Meanwhile, however, you might want to try the following script. Copy and paste the script into the AppleScript Editor's window and save it as an application. Then drop your screen capture files on the droplet's Finder icon.
    on open theDroppedFiles
        tell application "Finder"
            repeat with thisFile in theDroppedFiles
                set theFileName to name of thisFile
                if (word 1 of theFileName is "Picture") and ¬
                    (word 2 of theFileName = word -2 of theFileName) then
                    set theExtension to name extension of thisFile
                    set P to offset of theExtension in theFileName
                    set theCreationDate to creation date of thisFile
                    set dateStamp to short date string of theCreationDate
                    set timeStamp to time string of theCreationDate
                    set name of thisFile to "Screen Shot " & dateStamp ¬
                        & " at " & timeStamp & "." & theExtension
                end if
            end repeat
        end tell
    end open
    Message was edited by: Pierre L.

Maybe you are looking for

  • Restrict some users on Employee Vendor line items in FBL1N

    Dear All, We have a requirement in our current project. Employees have been created as vendors and they have specific number range. Client want to give access to see Emp Vendors data in FBL1N only to specific users. I need help on how to acheive this

  • Why won't my computer play a song that I just bought and my computer is already authorized

    I bought a a song off of iTunes and downloaded it.  Now every time I try to play the song I get a message saying that I need to authorize my computer.  However, when I authorize the computer, I get another message saying I have already authorized thi

  • Creating .msg file for MS Outlook

    Hi, I need to create a ".msg" file, so that some users can click on the message, and it opens up in MS Outlook. Has anybody successfully done this using the JavaMail APIs? Using the writeTo method (ie: message.writeTo(fileOut); ), I was able to creat

  • Obtaining monitor lock of a blocked thread

    Is there a way to obtain the monitor lock (reference to the object) of a blocked thread? Something like Object Thread.getMonitorLockedObject()Also when you get the locked object it would be useful if you can get reference to the thread which is holdi

  • I'm getting a message "An error occurred when attempting to change modules"

    Hi, I have just loaded LR CC on to my new iMac & I'm getting a message "An error occurred when attempting to change modules" I have used a Catalog from my Macbook stored on an external hard drive & there are no images, no panels & i can't access LR p