Maintaining Text data for IT0102 in Ad Hoc Query

We use the "Maintain Text" function to store comments and additional data for Grievance information on IT0102.  I need these comments to be included in a query for IT0102, however I only see the indicator for if text exists...there's no "field" to select to show the actual text - mainly because it's a note and not a field.
Any suggestions???

Hi,
you could make a custom field in your infoset for infotyp 0012. Then you could append code to this field and read the text with function module
HR_ECM_READ_TEXT_INFOTYPE into this field. 
Regards Bernd

Similar Messages

  • Maintain Validity Date for Documents

    Hi
    Any Idea how can we put and maintain Validity Dates for any Document uploaded in DMS Server.
    Validity Dates here means (Documents cannot be used if validty is over)
    I could not find any option in CV01*   T-Codes.
    Regards

    Dear Prasant,
    You can also do this solution also if you find it OK.
    1. Create a Class with "Validaty Date" as characteristics. Assign this Charac to class. Assign this class to doc Type. This charac should not be editable to any one except creator.
    2. Create a program with BAPI " Set Status" ( Check the name). which will run in background to check the validity date of Document. If the validity date is reached the BAPI will change the status to "S" as stated by Chritoph.
    This will help user to manually tracked which documents have exceded the validity date. You can also genegrate report and mail it to particular user that what documents have been set to Status "S".
    Hope I have provided a valid solution as I have tried similar kind of functionality with changing a characteristics with Date.
    With warm Regards
    Mangesh Pande

  • Maintain Validity Date for Mitigation Control Assignment to Users Virsa 5.2

    We have over 1,000 SoD's all mitigated.  The val;idity date for these mitigation controls needs to be updated.  Does anyone know a way to perform a range of updates so it is not necessary to update each user assigned to a Mitigation Control.

    The only way to do that currently would be to download the table information, edit in Excel and re-upload the table.
    Not for the faint of heart, but doable.
    Frank.

  • Maintain Shipping data for plants

    Dear all,
    We have Two plants which are of two different Sales Org. thru SD delivery, two step procedure.
    Path: Materials management> Purchasing>Purchase Order>Setup Stock transport Order>
    Define Shipping Data for Plant.
    Now the problem is here is i am able to assign only 1 Sales area for the corresponding depot. But
    we sell goods from both the plants via depot, the same way i want to send back the return goods
    from Depot to corresponding plants through STO. Please suggest
    Shipping point determination is maintaned properly but still getting problem..
    With Regards,
    San Rao.

    Hi,
    i  am unable to assign  two Sales org for the corresponding depot. it is having only one option.
    how can i assign second sales org.
    With regards,
    San Rao.

  • Is it too much to ask MS for a simple ad-hoc query tool for end users?

    Hello
    This is a general MS BI topic and not specific to Power Query, but I couldn't find a more suitable general forum for it.
    This posting is the result of the frustration I have with MS BI stack, whereby a simple ad-hoc drag and drop query tool to get the data in a table format (preferably from a relational
    data source) is missing, something that other big vendors have, such as SAP Business Objects Universe and the associated tools. The deprecated SSRS Report model and SMDL was the closest MS ever got to this functionality (although that had some limitations
    too), which unfortunately is dead now in favour of all sort of "Power" tools that basically can’t deliver the same functionality.
    The requirement is pretty simple: provide users with a tool that they can drag and drop fields from different entities in a data model without having to worry about entity relationships and
    apply necessary filters and get the data out in Excel or CSV, no fancy (or even simple) aggregation or formatting is required, and users should only see the rows they are authorized to see, i.e. row level security (RLS) of data.
    Let’s see what we have at our disposal from MS BI:
    From the modeling perspective, we have relational models in the database that support RLS with views, but obviously cannot maintain PK-FK relationships in the database model between views. Other
    models are .Net Entity Data model Framework model exposed as OData, and SSAS tabular, SSAS PowerPivot, SSAS multidimensional, which you can find a good comparison of their features here:
    http://www.sqlchick.com/entries/2012/3/4/decisions-powerpivot-ssas-tabular -or-ssas-multidimensional-m.html
    From the tools perspective we have:
    Power Query on top of a relational database (SQL Server): Can connect to relational databases easily, generated queries are optimal, detects relationships with entities, but only when entities
    are tables and have PK-FK relationships enforced in the database. So can’t detect relationships with views and hence can’t enforce RLS, so this is not an option unless users keep joining entities, which is not a desirable option either.
    Power Query on top of an OData: Works well except that queries generated for multiple entity expansion are terrible (look here for more details:
    http://social.technet.microsoft.com/Forums/en-US/7a8b8e52-d3a9-42af-bc52-9d5459c00529/odata-multilevel-expand-query-issue?forum=powerquery). In addition all filtering and projections are done at client side, so even if you want 1 row out of 1 million, all
    rows are brought over to Excel before filtered. Overall this results in unacceptable performance even for small sized data.
    Power Query on top of a tabular model: Doesn’t support that type of data source yet (http://social.technet.microsoft.com/Forums/en-US/ebcccb5e-7058-4046-917d-3d7fcf60cbdc/tabular-model-as-data-source-for-power-query?forum=powerquery).
    Even if it did, tabular models don’t support Many-to-Many relationships out of the box and not sure if there are plans to support tabular models in Direct Query mode in Power Query to take care of RLS.
    Excel on top of tabular models or multidimensional models: The tabular model should not be in Direct Query mode and you only get to use PivotTable to query the model. PivotTables are not famous
    for generating optimal MDX queries, so performance is not going to be acceptable if you are dealing with large dimensions, especially large junk and degenerate dimensions. You can use drilltrough to get details of a measure, but you have to specify the columns
    for a drilltrough action upfront, no flexibility for users to select what they want to see.
    PowerPivot on top of relational databases: The query designer cannot detect relationships between views, so users have to perform manual joins, not an option.
    PowerPivot on top of a multidimensional or tabular model: The query designer doesn’t generate optimal MDX queries by drag and drop. Especially there doesn’t seem to be a way to add related
    attributes of a dimension attribute to the query without cross joins which results in poor performance on large dimensions. Other issues found here
    http://sqlblog.com/blogs/marco_russo/archive/2010/08/24/using-ssas-2005-2008-as-powerpivot-data-source-query-designer.aspx. In addition the result set has ugly column names and issues with data types, such as interpreting numeric data types as Text, and
    there’s no straight forward way of getting the data from PowerPivot directly to Excel. The user either has to do a copy/paste (which doesn’t work on large datasets) or create a pivot table, bring the measure, and drillthrough the measure to get the details
    in table format.
    PowerPivot with OData: Treats every entity as a separate table to import, not suitable for our use case.
    Report Builder and SSRS on top of relational, multidimensional or tabular model: Given the query designer is the same as PowerPivot, it has similar issues. In addition, there are 4 more steps
    from designing a query to generating an Excel or CSV output, not an option.
    PowerView: Works against tabular and multidimensional models, but there’s no way of exporting data to Excel sheet. Besides the add-in is not available for Excel 2010.
    Good old MS-Access: Link to tables and views through ODBC, create relationships between entities and give each user a copy of that Access database to create their queries, and deal with the
    nightmares of updates to the model and distribution of new models to the users! Did I mention Access as an option?
    So there you go, 10 options, none of which can deal with a seemingly simple use case sufficiently. At the moment, the best among them appears to be option 6 against a heavily optimized OLAP
    cube, and perhaps #3 when/if Power Query support for tabular models in direct query mode arrives. Perhaps I am missing something very obvious, and would like someone to enlighten me here.
    Thanks
    MR

    Hi, as far as I can infer, that term means it's not loading any resources until needed, opposite of this...
    -bind_at_load
                     Sets a bit in the mach header of the resulting binary which tells dyld to bind all symbols                when the binary is loaded, rather than lazily.
    https://developer.apple.com/library/mac/documentation/Darwin/Reference/ManPages/ man1/ld.1.html
    Load resources lazily.
    You should never load a resource file until it is actually needed. Prefetching resource files may seem like a way to save time, but this practice actually slows down your app right away. In addition, if you end up not using the resource, loading it wastes memory for no good purpose.
    http://developer.apple.com/library/ios/#documentation/iphone/conceptual/iphoneos programmingguide/PerformanceTuning/PerformanceTuning.html
    And, higher OS versionsare even worse with Bonjour/mDNSResponder, but might see this...
    http://support.apple.com/kb/ht3789

  • Where to find 'Exempt' indicator for reporting using ad-hoc query?

    I want to run report for a count of exempt and non-exempt employees using ad-hoc query with logical database PNPCE. We maintain the exempt/non-exempt flag on the job.
    Thanks
    Manish

    To meet this exact same requirement at my company.  I added an additional table (T5U13) into the logical database. Within transaction SQ02, go to the field group Organizational Assignment, right-click on the field group.  Click Node Extras.  Click Create icon over on the right-hand side of the screen.  Type in the table name T5U13.  Use this select statement:
    SELECT SINGLE * FROM T5U13              
      WHERE STELL                = P0001-STELL
        AND ENDDA                = P0001-ENDDA
    Then add the fields into your field groups.  I also created a local field so I get the word "Exempt" or "Non-Exempt" in my queries instead of "1" or "2" because i can never remember which is which. 
    Hope this helps!

  • Text data for condition type  & customer po number

    Will there be any text datsrc available for  condition type  & customer po number
    Example I am displaying condition type data & customer po number data in my report
    condition typ    customer po nu
    JASS                        1001
    JEEP                         121

    Hi,
    goto Tx LBWE, application 11, maintain extract structure MC11VA0HR.
    Check if your field is on the right side, you should find at least the following
    MCVBAK  CMFRE release date
    MCVBAK  KTEXT Description
    if they're on the right pane, then you'll have to include them for extraction; depending on if the delta is already initialez or not, if you want past history as well, the procedure might be more less complicated.
    regards
    Olivier.

  • Load Same Text data for all language key

    Hello Everyone,
    I have a datasource with Material and Desc fields but no language key.I need to load the same desc into my infoobject for all the languages i am maintaining (EN/DE/IT/FR/PL). Is this possible?
    Thanks,
    Ajay.D

    Hi Ajay,
    you can load same description for all the languages by maintaining constant language value at Transfer rules/ Transformations.
    but here you need to change the constant value for each language.
    Ex, if you are loading for EN, maintain language as Constant "EN".
    and then for DE, maintain language as Constant "DE".
    but this is repetitive process, if new materials are added then need to load again.
    Best Regards.
    Edited by: Daya Sagar on Sep 16, 2009 11:16 AM

  • Problems with maintaining master data for infoobject 0SOLD_TO

    Hello experts,
    I'm not getting 'maintain master dada' option when I right click on the infoobject 0SOLD_TO. The data is already there from pervious data loads (from R/3 systems) and is visible in the reports. I'm using this infoobject in cube 0SD_C05 to report quotation/order details. Am I missing something here? Please guide.
    Thanks
    Sumit

    Hi Sumit,
    It looks you have asked your question in a wrong forum.
    Please use BI forum to post your question.
    Regards,
    Pooja

  • Infoobject change for extracting texts data.

    Hi BW guys,
    Here is my requirement.
    I have one info object 'salesmen', which is already used in some other ODS & Cube's.
    Now I want to extract texts data for the object 'salesmen', for that I will need to change my infoobject (changes are : adding credit control are object under compounding).
    But while i am activating the info object again it is giving errors.
    Error messages:
    1) InfoObject XXXXX (or ref.) is used in data targets with data -> Error:
    2) Characteristic XXXXX: Compound or reference was changed
    3)InfoObject XXXXX being used in InfoCube XXXX (contains data)
    etc....
    But i don't want to delete the data in any data target.
    Is there any way to solve this problem?
    Thanks in advance......

    Hi,
    If you have not many cubes and ODSs with this salesman, you can consider another, beter, but more time-consuming way.
    1. Create a new IO for your salesman, add a compounding attribute as you want.
    2. Load master data for the new IO.
    3. Create copies of your infoproviders.
    3. In each of them delete an old salesman IO and insert a new one.
    4. Create export datasourses for old cubes.
    5. Create update rules for new data targets based on old ones.
    6. In URs map your new IO with the old one. All other IOs should be mapped 1:1 (new<-old).
    7. Reload data targets.
    That's all.
    The way I proposed earlier is less preferrable. Because anyway you'll have to change loaded into data targets data. And in this case it's better to change data model as you want.
    Best regards,
    Eugene

  • Need Text data table for GBSTA field(ref VBUP table)

    Hi frns,
    Can anyone suggest me the table to extract text data for the field GBSTA .
    Cheers,
    Shana

    Hi,
    The value range of domain "STATV" for field "GBSTA" is limited with fixed values. You can either read entries from Value range of domain "GBSTA" in SE11 or from table TVBST
    Table : TVBST with inputs
    SPRAS : sy-langu
    TBNAM : VBUP
    FDNAM : GBSTA
    Srikanth.

  • Text ID for text object BUT000 does not exist

    Hi Experts.
    I am doing data migration using LSMW with IDoc. I have this error message "Text ID for text object BUT000 does not exist" when processing the IDoc.
    Can anyone tell me what is this error about and how could i fix it. A step by step guide would be nice. Really need advice urgently.
    I'll definitely rewards points.
    Regards,
    Julianna

    for this u need to maintain the text object and its related text id using the TC: SM31 in rel. view of  ur tcode( like MM01, Va01 etc).
    first get the view names(1 for text object and 2 for text id) from fun. people.
    then go to sm31 with 1 view, create BUT000 .
    then go to sm31 with 2 view, maintain text id for the text object:BUT000.

  • Unable to see text data in the reporting bw 3.5

    Dear Gurus,
                         My clinet wants to see the text data for all the records in the reporting.unfortunately he cant see text data for only 5 records in the colum. i checked in the query designer for that object ,in properties its given as key and text. can any one suggest me why the text data is not comming for only that records..
    Thanks in Advance.
    Srikanth.

    Dear Pavan,
                            Thanks for the reply, the query properties are already set like key and text before, the problem text is showing to all the records except 6 records,for this 6 records in the place of text it is showing only same technical name..please guide..
    Thanks,
    srikanth.

  • Error :Maintain pricing conditions for the material for the excise invoice

    Hi
    i have maintained excise data for material, after that i have created scheduled agreement with lp and with 0 value private and dummy tax code  and delivery schedule is released. Now when i do the goods receipt i am getting error " Maintain pricing conditions for the material for the excise invoice date
    Message no. 8I629" can some body can help me.
    with regards
    Narendra kumar

    Hi,
    goto tcode fv11 u have to maintained the condition.
    first enter the condition type like jmop -
    then enter, next screen u have the select the below comination.
    select the plant/vendor/material base u have to maintain the conditions.
    Thank's
    MVS

  • Master text data in Open Hub

    Hi All,
    I am working on BI 7.0 and am currently working on Open Hub. I have to extract data from a cost center infocube, but i also need to send the master text data for certain fields along with the infocube data into the file server.
    Is this possible with open hub services,could someone please provide me with a step by step procedure to this one.
    I have been following a similar thread
    Nav. Attrib in Open Hub 04s
    Please let me know if it was succefully implemented by someone.
    Regards
    Dipali

    Hi Uppala,
    Following are 2 approaches to achieve this
    Approach 1:
    1. Create info spoke which will extract data from cube to data base table (do not use file)
    2. Create one more info spoke which will extract master data text to data base table (again do not use file)
    3. Create join on both the data base tables.
    4. Write a ABAP program which will create a file based on the joins of tables.
    Approach 2 :
    1. Create multiprovider on cube and master data info object
    2. Create info spoke based on the multi provider, that will extract data to flat file
    For 2nd approach u have to put certain filters to avoid unwanted rows
    Regards,
    Ajinkya

Maybe you are looking for