Urgent regarding Data & Task Audits Back Up

Hi Experts
Here my question is regarding Data Audit and Task audits in V 11.1.2.1
-->If we take the back up of Data & Task audit tables data in to other backup_tables in same schema how this can be viewed through task audit and data audit through HFM?
-->Whether any third part tools available to audit the tasks (DATA & TASK ) information...?
-->Whether EPM maestro can be used here..?
Thanks in advance
Edited by: RajaKK on May 29, 2012 6:19 PM

Use the Audit Extract utility bundled with HFM 11.1.2.1 to export this information periodically to a CSV file that you can view offline through any text editor, or even Excel. There is a command line feature for this as well, so you could incorporate this into a batch routine.
For anyone attending Kaleidoscope in San Antonio, TX next month, I will present the various utilities that ship with HFM including this one. Hope to see you there!
--Chris                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                               

Similar Messages

  • URGENT REGARDING  DATE

    hI,
    IN SELECTION SCREEN DEFINED
    SELECT-OPTIONS :  p_budat FOR mkpf-budat.
    How can I check these dates can anybody give coding part to this.
    Regards
    SAISRI

    Hi
         Your query is not very Clear.. Anyway if you want to check the dates that is coming in this interval Do like This......
    SELECT-OPTIONS : p_budat FOR mkpf-budat.
    data:begin of it_mkpf occurs 10,
           budat type bkpf-budat,
           end if it_mkpf.
    select budat from mkpf inti it_mkpf where budat in p_budat.
    Now you can check your condition in internal table..
    If this is not the exact requirment get back..
    Reward All Helpfull Answers.............

  • Urgent:Regarding data reconsilation

    HI all,
    1. i have PSFT as trusted source and EBS as target source. now when we connect to this sources by using out-of-the-box connectors which is already avilable,what it will ask while installing connectors and which tables it is going to connect?
    2.i have a requirement that i need get the data related to users from diff tables of PSFT and need to provision that data in diff tables of EBS,wether we need to fulfill this by using predefind connectors or we need to use custom connetors?
    if possible please give me the steps for the above queries.
    please help me out in this
    thanks for your time
    Regards
    siva

    PSFT needs profile to be set at PSFT for the data and EBS has queries in the properties file with the connector. Installation does not in itself does anything but you will have to configure the connectors post installation as per your requirement. My suggestion would be to go through the connector docs to understand the architecture and feasibility.
    HTH,
    BB

  • URGENT: regarding data load in essbase through ODI after upgradation

    Hi,
    I have got this error, Please give me the solution...
    org.apache.bsf.BSFException: exception from Jython:
    Traceback (most recent call last):
    File "<string>", line 29, in <module>
         at org.hsqldb.jdbc.Util.sqlException(Unknown Source)
         at org.hsqldb.jdbc.JDBCStatement.fetchResult(Unknown Source)
         at org.hsqldb.jdbc.JDBCStatement.executeQuery(Unknown Source)
         at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
         at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
         at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
         at java.lang.reflect.Method.invoke(Method.java:597)
    java.sql.SQLException: java.sql.SQLException: unexpected token: -
         at org.apache.bsf.engines.jython.JythonEngine.exec(JythonEngine.java:146)
         at com.sunopsis.dwg.codeinterpretor.SnpScriptingInterpretor.execInBSFEngine(SnpScriptingInterpretor.java:346)
         at com.sunopsis.dwg.codeinterpretor.SnpScriptingInterpretor.exec(SnpScriptingInterpretor.java:170)
         at com.sunopsis.dwg.dbobj.SnpSessTaskSql.scripting(SnpSessTaskSql.java:2458)
         at oracle.odi.runtime.agent.execution.cmd.ScriptingExecutor.execute(ScriptingExecutor.java:48)
         at oracle.odi.runtime.agent.execution.cmd.ScriptingExecutor.execute(ScriptingExecutor.java:1)
         at oracle.odi.runtime.agent.execution.TaskExecutionHandler.handleTask(TaskExecutionHandler.java:50)
         at com.sunopsis.dwg.dbobj.SnpSessTaskSql.processTask(SnpSessTaskSql.java:2906)
         at com.sunopsis.dwg.dbobj.SnpSessTaskSql.treatTask(SnpSessTaskSql.java:2609)
         at com.sunopsis.dwg.dbobj.SnpSessStep.treatAttachedTasks(SnpSessStep.java:540)
         at com.sunopsis.dwg.dbobj.SnpSessStep.treatSessStep(SnpSessStep.java:453)
         at com.sunopsis.dwg.dbobj.SnpSession.treatSession(SnpSession.java:1740)
         at oracle.odi.runtime.agent.processor.impl.StartSessRequestProcessor$2.doAction(StartSessRequestProcessor.java:338)
         at oracle.odi.core.persistence.dwgobject.DwgObjectTemplate.execute(DwgObjectTemplate.java:214)
         at oracle.odi.runtime.agent.processor.impl.StartSessRequestProcessor.doProcessStartSessTask(StartSessRequestProcessor.java:272)
         at oracle.odi.runtime.agent.processor.impl.StartSessRequestProcessor.access$0(StartSessRequestProcessor.java:263)
         at oracle.odi.runtime.agent.processor.impl.StartSessRequestProcessor$StartSessTask.doExecute(StartSessRequestProcessor.java:822)
         at oracle.odi.runtime.agent.processor.task.AgentTask.execute(AgentTask.java:123)
         at oracle.odi.runtime.agent.support.DefaultAgentTaskExecutor$2.run(DefaultAgentTaskExecutor.java:83)
         at java.lang.Thread.run(Thread.java:662)
    Thanks,
    Rubi

    I have used some condition in data column.
    If I have deleted that condition and directly mapped in data column them interface works fine.
    But whenever I have used some normal CONVERT function like CONVERT("Source column name",Numeric), It gives me that error.

  • URGENT: REGARDING DATE DISPLAY IN REPORT

    hi,
    i am making a report in which i have to display that these are the persons who had released the requsitions at these dates,i am able to display the persons who had released the requsitions but i am not able to display the dates on which they had released the requsitions.
    if any body who  will provide me help for this solution he or she will be defiantely rewarded...

    hi ric,
    i also tested and it gave this o/p.
    28.11.2007                                                                               
    PROGRAM FOR TESTING                                                                               
    1
    PUR.REQ   DATE    REQ.TYPE   1ST APPROVAL  2ND APPROVAL  3RD APPROVAL
    PRESON       DATE      PERSON      DATE     PERSON     DATE
    10009603   19.07.2005 NB   A1             19.07.2005
    10010046   08.03.2006 NB   R1 R2          08.03.2006
    10010115   27.03.2006 NB   EX             27.03.2006
    10010116   27.03.2006 NB   EX             27.03.2006
    10010827   22.05.2006 NB   G1 G2          22.05.2006
    i think it is right...
    what else you need to display?

  • Analysing Task Audit, Data Audit and Process Flow History

    Hi,
    Internal Audit dept has requested a bunch of information, that we need to compile from Task Audit, Data Audit and Process Flow History logs. We do have all the info available, however not in a format that allows proper "reporting" of log information. What is the best way to handle HFM logs so that we can quickly filter and export required audit information?
    We do have housekeeping in place, so the logs are partial "live" db tables, and partial purged tables that were exported to Excel to archive the historical log info.
    Many Thanks.

    I thought I posted this Friday, but I just noticed I never hit the 'Post Message Button', ha ha.
    This info below will help you translate some of the information in the tables, etc. You could report on it from the Audit tables directly or move them to another appropriate data table for analysis later. The concensus, though I disagree, is that you will suffer performance issues if your audit tables get too big, so you want to move them periodically. You can do this using a scheduled Task, manual process, etc.
    I personally just dump it to another table and report on it from there. As mentioned above, you'll need to translate some of the information as it is not 'human readable' in the database.
    For instance, if I wanted to pull Metadata Load, Rules Load, Member List load, you could run a query like this. (NOTE: strAppName should be equal to the name of your application .... )
    The main tricks to know at least for task audit table are figuring out how to convert times and determing which activity code corresponds to the user friendly name.
    -- Declare working variables --
    declare @dtStartDate as nvarchar(20)
    declare @dtEndDate as nvarchar(20)
    declare @strAppName as nvarchar(20)
    declare @strSQL as nvarchar(4000)
    -- Initialize working variables --
    set @dtStartDate = '1/1/2012'
    set @dtEndDate = '8/31/2012'
    set @strAppName = 'YourAppNameHere'
    --Get Rules Load, Metadata, Member List
    set @strSQL = '
    select sUserName as "User", ''Rules Load'' as Activity, cast(StartTime-2 as smalldatetime) as "Time Start",
          cast(EndTime-2 as smalldatetime) as ''Time End'', ServerName, strDescription, strModuleName
       from ' + @strAppName + '_task_audit ta, hsv_activity_users au
       where au.lUserID = ta.ActivityUserID and activitycode in (1)
            and cast(StartTime-2 as smalldatetime) between ''' + @dtStartDate + ''' and ''' + @dtEndDate + '''
    union all
    select sUserName as "User", ''Metadata Load'' as Activity, cast(StartTime-2 as smalldatetime) as "Time Start",
          cast(EndTime-2 as smalldatetime) as ''Time End'', ServerName, strDescription, strModuleName
       from ' + @strAppName + '_task_audit ta, hsv_activity_users au
       where au.lUserID = ta.ActivityUserID and activitycode in (21)
            and cast(StartTime-2 as smalldatetime) between ''' + @dtStartDate + ''' and ''' + @dtEndDate + '''
    union all
    select sUserName as "User", ''Memberlist Load'' as Activity, cast(StartTime-2 as smalldatetime) as "Time Start",
          cast(EndTime-2 as smalldatetime) as ''Time End'', ServerName, strDescription, strModuleName
       from ' + @strAppName + '_task_audit ta, hsv_activity_users au
       where au.lUserID = ta.ActivityUserID and activitycode in (23)
            and cast(StartTime-2 as smalldatetime) between ''' + @dtStartDate + ''' and ''' + @dtEndDate + ''''
    exec sp_executesql @strSQLIn regards to activity codes, here's a quick breakdown on those ....
    ActivityID     ActivityName
    0     Idle
    1     Rules Load
    2     Rules Scan
    3     Rules Extract
    4     Consolidation
    5     Chart Logic
    6     Translation
    7     Custom Logic
    8     Allocate
    9     Data Load
    10     Data Extract
    11     Data Extract via HAL
    12     Data Entry
    13     Data Retrieval
    14     Data Clear
    15     Data Copy
    16     Journal Entry
    17     Journal Retrieval
    18     Journal Posting
    19     Journal Unposting
    20     Journal Template Entry
    21     Metadata Load
    22     Metadata Extract
    23     Member List Load
    24     Member List Scan
    25     Member List Extract
    26     Security Load
    27     Security Scan
    28     Security Extract
    29     Logon
    30     Logon Failure
    31     Logoff
    32     External
    33     Metadata Scan
    34     Data Scan
    35     Extended Analytics Export
    36     Extended Analytics Schema Delete
    37     Transactions Load
    38     Transactions Extract
    39     Document Attachments
    40     Document Detachments
    41     Create Transactions
    42     Edit Transactions
    43     Delete Transactions
    44     Post Transactions
    45     Unpost Transactions
    46     Delete Invalid Records
    47     Data Audit Purged
    48     Task Audit Purged
    49     Post All Transactions
    50     Unpost All Transactions
    51     Delete All Transactions
    52     Unmatch All Transactions
    53     Auto Match by ID
    54     Auto Match by Account
    55     Intercompany Matching Report by ID
    56     Intercompany Matching Report by Acct
    57     Intercompany Transaction Report
    58     Manual Match
    59     Unmatch Selected
    60     Manage IC Periods
    61     Lock/Unlock IC Entities
    62     Manage IC Reason Codes
    63     Null

  • Data Retrieval in task audit

    Hello,
    What is the exact meaning of Data Retrieval in task audit? Is it just data retrieved by reports?
    Thanks
    Lilach

    Hi Lilach,
    It would mean data retrieval through adhoc analysis or grids,HsGetValue Functions etc.
    You could also check Task Audit table you can see activity codes which has details below.
    ActivityID
    ActivityName
    0
    Rules Load
    1
    Rules Scan
    2
    Rules Extract
    3
    Consolidation 
    4
    Chart Logic
    5
    Translation 
    6
    Custom Logic
    7
    Allocate 
    8
    Data Load
    9
    Data Extract
    10
    Data Extract via HAL
    11
    Data Entry
    12
    Data Retrieval
    13
    Data Clear
    14
    Data Copy
    15
    Journal Entry
    16
    Journal Retrieval
    17
    Journal Posting
    18
    Journal Unposting
    You can query AppName_Task_Audit & HSV_Activity_Users tables by joining them and applying condition on Activity id from App_Task_Audit.
    Thanks
    Amith

  • How to guarantee that all events regarding Data Cache are dispatched when application is terminating? Urgent

    Hello,
    we have the phenomena that when an application is commiting a transaction and then terminating,
    often not all events regarding Data Cache are dispatched by the TCP kodo.RemoteCommitProvider.
    It seems that the JVM on termination is not waiting until RemoteCommitProvider has dispatched all events regarding Data
    Cache. In this way we sometimes loose some cache synchronization and some of our customers run into serious problems.
    Is there a way to guarantee that all Data Cache events are dispatched before the aplciation terminates
    (maybe implementing a shutdown hook?).
    best regards
    Guenther Demetz
    Wuerth-Phoenix SRL

    Hi,
    as nobody answered to my question I try to explain it more simple:
    Are the TCP-kodo.RemoteCommitProvider threads acting as user threads or as threads of type 'deamon' ?
    I hope that soon someon can answer
    best regards
    Guenther Demetz
    Wuerth-Phoenix SRL
    G.Demetz wrote:
    Hello,
    we have the phenomena that when an application is commiting a transaction and then terminating,
    often not all events regarding Data Cache are dispatched by the TCP kodo.RemoteCommitProvider.
    It seems that the JVM on termination is not waiting until RemoteCommitProvider has dispatched all events regarding Data
    Cache. In this way we sometimes loose some cache synchronization and some of our customers run into serious problems.
    Is there a way to guarantee that all Data Cache events are dispatched before the aplciation terminates
    (maybe implementing a shutdown hook?).
    best regards
    Guenther Demetz
    Wuerth-Phoenix SRL

  • Automate HFM Task Audit Export

    I am trying to find a way to export the daily task audit info and email it to our auditors on a nightly basis. I figure if I can find out which table in the HFM database contains this audit info I can write something to automatically export it. Does anyone know where I can find this audit task data? I have searched tables in the HFM database and I am unable to determine this.
    Thanks in advance
    HFM Version 11.1.2.2
    DB SQL Server 2008 R2

    Hi there,
    Check the Consultant Tools that are installed with HFM. There is a quite simple tool which allows you to extract the task and data audit without any effort at all.
    Regards,
    Thanos

  • Is there a major flaw with Iphoto in regards to organizing and backing up?

    Is there a major flaw with Iphoto in regards to organizing and backing up? And how should I proceed??
    Here is my story. I have moved roughly 15,000 photos from 7 computers (4 PCs and 3 Macs - basically everyone's computer in the family) onto one external hard drive. The photos were taken by many different digital cameras, are sparsely labeled, and have the most random names and folders. There are a LOT of duplicates. MY MISSION - label and back them up.
    The way I normally deal with photos/DATA is this: I plug my camera into my Mac and immediately think about organizing and backing them up. With "Preview" I look at photos (and decide which one I like), create a folder called "where-ever I took these photos" and drag them into that folder. The problem with Iphoto is that you have no ability to control folders, unlike many other applications.
    For example, in Itunes you can rename an artist, or even an album, and it will automatically appear in an updated folder. If you alter a photo in Photoshop, it makes you save it is a new file. In Iphoto you cannot make the same "root level" changes without making duplicates. From what I understand, you can place photos into an Album and then export that album (which makes duplicates). Iphoto will rename each duplicate file and then place it in the correct folder. The problem here is I have no idea where the original is to delete it.
    Basically I want to create folders called "Trip to Italy 2003" or "Rachel's Wedding" and place all the photos associated with that event in those folders. The problem I run into is that I have a bunch of duplicates. I have heard of a program called Duplicate Annihilator which has gotten favorable reviews. If I import them into Iphoto all the photos are placed in random named folders. I say random because some of the photos were taken on cameras that didn't have the date set correctly. Hence some of my photos will appear in a labyrinth of folders called "Modified / 2001 / Roll 17 / DCM482 / Image25874654.jpg"
    I am totally willing to wipe my Iphoto library clean and start again if there is a way to organize the photos as described above. Eventually I will take these folders and save them on a data DVD.
    Thanks to anyone who can help.
    Gary

    You do not need to delve into the folders within the iPhoto Library via the finder to access any photo. That is all done within iPhoto itself. If you need a photo for use outside of iPhoto read Terence Devlin's treatise on file access. iPhoto is a DAM application - Digital Asset Management, not a folder manager. There's a big difference. You can use keywords, album and smart albums to organize your photos.
    If you still have the original files you might consider what I do. I rename each file with the date it was taken using the international date format: YYYY-MM-DD-001.jpg. You can download the demo version ofMedia Expression and automatically rename each folder of photos with the EXIF date and a sequential number. This makes chronological searching and sorting much easier both in iPhoto and the Finder.
    If you want to not only manage the photos but the folders then Expression Media will let you do both. It will let you rename files after being imported into a catalog, move them between folders, add ITPC metadata to the files, create virtual albums called Supplemental categories, and much, much more. I use it as my primary DAM and iPhoto for special projects like books, slideshows, calendars, etc.

  • Export Task Audit Not Working

    Hello,
    We are on HFM 11.1.2.1.103 and I am getting an error message when trying to export a users activity from the "Task Audit" menu.  When I click on export a new tab comes up with the URL http://server:port/hfm/Administration/TaskAuditExport.asp and then I get an error message:
    An error occurred on the server when processing the URL. Please contact the
    system administrator.
    If you are the system administrator please click here to find out more
    about this error.
    And when I click on "here" it takes me to this URL Running Classic ASP Applications on IIS 7 and IIS 8 : The Official Microsoft IIS Site
    Any thoughts on how I resolve this would be much appreciated.
    Thank you,
    Jason

    SDM,
    Thank you for the response.  This is helpful and I was able to extrac the data, but this utility doesn't give you the ability to focus on one users' activity as you can from the web.  Why is it not recommended to use the web and is there a way to get that working properly?
    The reason the utility is not working for my purpose is I want to view a few users' activities over the last year and when I run the utility it gives me everyone and extracts the data across numerous files, which is not particularly easy to work with.
    Thanks.

  • Regarding date types

    hi all,
    i need some information regarding date types infotype 41.
    i need to print date types in sequential manner like dar01 dar02 dar03 like upto 12 date types.
    can you please suggest me how should i print them.
    thanks,
    madhavi.

    Hi,
    Steps to create a HR Infotype:
    1) Go to Transaction PM01.
    2) Enter the custom Infotype number which you want to create (Should be a 4 digit number, start with 9).
    3) Select the 'Employee Infotype' radio button.
    4) Select the 'PS Structure Infotype'.
    5) Click on Create... A separate table maintenance window appears...
    6) Create a PS structure with all the fields you want on the Infotype
    7) Save and Activate the PS structure
    8) Go back to the initial screen of PM01.
    9) Click on 'All' push button. It takes a few moments.
    10) Click on 'Technical Characteristics'. Infotype list screen appears
    11) Click on 'Change'(pencil) button
    12) Select your Infotype and click on 'Detail' (magnifying glass) button
    13) Give 'T591A' as subtype table
    14) Give 'T591S' as subtype txt tab
    15) Give your subtype field as subtype field
    16) Save and come back to PM01 initial screen
    17) Click on 'Infotype Characteristics' ... Infotype list screen appears
    18) Click on 'Change' (pencil) button
    19) Click on 'New Entries'
    20) Enter your Infotype number and short text
    21) Here we have to set different Infotype Characteristics as per the requirement. (Better open another session with some standard Infotype's infotype characteristics screen and use as the reference to fill yours)
    22) Save your entries.
    23) Now the Infotype is created and ready to use.
    24) If you want to change the layout of the Infotype as per your requirement...
    25) In the PM01 initial screen...Select 'Screen' radio button and give 2000 as the screen name, then click on edit.
    26) In the next screen.. Select 'Layout Editor' and click 'Change'.
    27) Screen default layout appears...here you can design/modify the screen..change the attributes of the fields..etc.
    28) Save and activate. (Don't forget to 'Activate at every level)
    Subtype Creation :
    Transaction PM01 Goto Subtype Characteristics. Click on Append and then subtype. Enter the name and description of subtype on screen.
    Then goto technical Characteristics and maintain the details of subtype there. I.e name of subtype i.e. component name defined in PSnnnn. Subtype table is T591A.
    Subty.text tab is T591S and time const tab is T591A.
    See:
    http://help.sap.com/saphelp_46c/helpdata/en/4f/d5268a575e11d189270000e8322f96/content.htm
    HR related site:
    http://www.sapdevelopment.co.uk/hr/hrhome.htm
    Enhancement of Infotype
    Check the following
    http://help.sap.com/printdocu/core/Print46c/en/data/pdf/PAXX/PYINT_INFOTYP.pdf
    Infotype Enhancement overview screen
    https://www.sdn.sap.com/irj/sdn/go/portal/prtroot/docs/library/uuid/60a7586d-edd9-2910-68a8-8204303835a1
    with regards,
    Hema SUndara.

  • I want to Subtract the current date to find out whats the date 10 days back

    Hi guys,
    I want to subtract the current date to find the date 10 days back. If I use Gregorian calendar and then substract the day
    calender.get(GregorianCalendar.DATE) - 10;
    I will get right if my current date is 20 or more. If it happens to be less than 10, then I get in negative.
    Please suggest an solution for this.
    Thanks in advance
    With regards,
    Amin

    Here's the code.
        GreogrianCalendar mycalendar = new GregorianCalendar() ;
        mycalendar.add(mycalendar.DATE, -10) ;That should get you the date you want and just ignore the background noise that is hambone.
    Hi guys,
    I want to subtract the current date to find the date
    10 days back. If I use Gregorian calendar and then
    substract the day
    calender.get(GregorianCalendar.DATE) - 10;
    I will get right if my current date is 20 or more. If
    it happens to be less than 10, then I get in negative.
    Please suggest an solution for this.
    Thanks in advance
    With regards,
    Amin

  • 10g parameters for Task Audit System 9

    Does anyone have suggestions regarding how to optimize for USER I/O during a metadata load the TASK_AUDIT table on 10g?
    The ADDM report suggests Segment Tuning and Application tuning (Application tuning not possible since it is HFM code.) Looking at a relatively large application, the database improvement would be about 10 minutes if we could minimize the USER I/O waits for this table.
    If you do not have a database recommendation, does anyone know if I can stop the logging of the task audit until metadata load completes?
    Thanks,

    We daily backup this table to a versionning table, and truncate it, it is much smaller and than the performance impact is much less
    do you need to keep the whole thing??

  • Standard Inport Package is getting failed at Convert Data Task

    Hi All,
    When running standard import package on any application, it's getting failed at Convert Data Task.
    Do any body faced similar kind of issue, please let me know what might be the cause or any solution to the above issue would be appreciated.
    Regards,
    Sreekanth.

    Hi,
    The format of dimension members in the external file is different from the BPC internal format. Check your conversion file. The problem is with your conversion file.
    For example:
    In the external file Time dimension values is given as Jan, Feb, Mar etc..
    In BPC asume you are maintaining time dimension members as 2008.Jan, 2008.Feb etc.
    Then you should define this in conversion file. Jan should be replaced by 2008.Jan when it enter BPC application. Also, the values you define in the Internal format should be maintained as master data (dimension members).
    regards,
    Lokesh Nandula

Maybe you are looking for