Large number of attributes for a characteristic

I have the requirement to add like 100 attributes to the 0Material object. This object already have 120 attributes, what would result in a total ol 220 attributes. Although I know the technical limit is 750, I'd like to know if anybody has heard or experienced some performance or any kind of issue because of having such a big number of attributes in an object (0Material being particularly critical). Thanks a bunch. Regards.

I dont know the upper limit, but I think 80 is too much. I have never used more than 15-20. For Nav attributes, a seperate tables are created which causes the Performance issue as result in new join at query run time. Just ask your business guy, if these can be reduced.One way could be to model these attributes as seperate characteristics. It will certainly help.
Thanks...
Shambhu

Similar Messages

  • Large number of attributes reduces data display in Query (PLS HELP).

    Dear Guru's
    We have newly upgraded our system to 7.0, Currently i have a Query based on master data info object(0 equipment) which contain several attributes, due to huge number of attributes, i cannot see data for few  attributes( Ex:-0service descrip), but data exists for these atrrib, when i check in RSA1. my client want to use heavy number of 0Euipment numbers(nearly 10000) in selection screen. thats the reason for data not to be seen.
    WE are under SP 16 also(Note 562162 - Additional selection of attribute reduces query result set is allready exists), and did not found any proper note.
    So guru's can any one please tell me how i can get the data in Query for these attribute by keeping the same selection option for Equipment numbers (10000).
    Dude's give me some good suggestion please...client is waiting from 1 week
    Thanks in adv
    Dev
    Edited by: Srinivas dev on Jun 10, 2008 11:52 AM

    Dear Yogesh,
    Thanks a lot dude...
    I assigned points for your kindly help....
    Got any idea about by initial post   Large number of attributes reduces data display in Query
    Thanks and Regards
    Dev

  • Best way to handle large number video files for a project..

    Hey, I was looking at getting some insight from the community here. Bascially there is a project that is being worked on that requires large amount of footage to be sifted through that only a small percentage will be used. These are mostly HD files and while most of the footage has been watched on Quicktime with notes taken, my question is this.
    What is the best way to take only small portions of each file without having to load everything into final cut and without any loose of quality. Should I just trim and rename from Quicktime or is there an easier way?
    Reason this needs to be done this way is the smaller segments will each be sent to other editors and rather then send huge files we want to split it to smaller amounts for each editor to use.
    Thank you so much for any input regarding this, I look forward to what you have to say

    Open the clip into the viewer. Mark In and Out points on the section you want. Make it a subclip Cmd-U. Drag the subclip into the bin for the editor who needs it. Repeat.
    If you batch export from a clip there is a selection to choose whether to export the whole clip or check box to export the marked I/O.
    This does not sound like a good project on which to being learning FCP.

  • Creating keyfigure as an attribute for a characteristic

    Hi All,
      Can anybody explain what is the use of creation of keyfigure as an attribute in the attributes creation process.Please provide a sample example as to where it i used in real time.
    Regards,
    Sunitha

    Hi ,
    Cost Price will still be defined as a key figure, but we will be defining it as an attribute of 0MATERIAL.
    Choice of doing this will subject to specific reporting requirements like, since cost price of the article is slowly changing value (nearly fixed) ,hence we might prefer to load this  into the master data which can be looked up along with other attributes of 0MATERIAL while doing transactional loads.
    Also this will help in reducing the width of the fact table.
    Another scenario could be if price is  time dependant. It would be a good option again to use this as an attribute.
    Thanks,
    Vikram

  • Using Key figure attribute of a characteristic in a query

    Hi All,
    I have the following scenario. I have defined a key figure attribute for a characteristic Infoobject. I have included this characteristic infoobject in an Infocube. I now want to use the key figure attribute as a regular key figure when I create a report based on the Infocube. Is it possible and if so how is it possible?
    Thanks
    Sundar

    Hi Saundar,
       There's another way.Set the characteristic with key figure attributes as a data target under your info area. Create a MultiProvider and include your Cube and this characteristic. Here you can set the key figure as your MultiProvider key figure.
       Don't forget to take care of the identification of MultiProvider.
       Hope this helps.
    Regards,
    Aaron

  • Rman backup failure, and is generating a large number of files.

    I would appreciate some pointers on this if possible, as I'm a bit of an rman novice.
    Our rman backup logs indicated a failure and in the directory where it puts its files there appeared a large number of files for the 18th, which was the date of the failure. Previous days backups generated 5 files of moderate size. When it failed it generated between 30 - 40 G of files ( it looks like one for each database file ).
    The full backup is early monday morning, and the rest incremental :
    I have placed the rman log, the script and a the full directory file listing here : http://www.tinshed.plus.com/rman/
    Thanks in advance - George
    -rw-r----- 1 oracle dba 1073750016 Jan 18 00:03 database_f734055071_s244_s1
    -rw-r----- 1 oracle dba 1073750016 Jan 18 00:03 database_f734055096_s245_s1
    -rw-r----- 1 oracle dba 1073750016 Jan 18 00:03 database_f734573008_s281_s1
    -rw-r----- 1 oracle dba 1073750016 Jan 18 00:03 database_f734055045_s243_s1
    -rw-r----- 1 oracle dba 524296192 Jan 18 00:03 database_f734055121_s246_s1
    -rw-r----- 1 oracle dba 1073750016 Jan 18 00:03 database_f734055020_s242_s1
    -rw-r----- 1 oracle dba 4294975488 Jan 18 00:02 database_f734054454_s233_s1
    -rw-r----- 1 oracle dba 4294975488 Jan 18 00:02 database_f734054519_s234_s1
    -rw-r----- 1 oracle dba 4294975488 Jan 18 00:02 database_f734054595_s235_s1
    -rw-r----- 1 oracle dba 4294975488 Jan 18 00:02 database_f734054660_s236_s1
    -rw-r----- 1 oracle dba 4294975488 Jan 18 00:02 database_f734054725_s237_s1
    -rw-r----- 1 oracle dba 4294975488 Jan 18 00:02 database_f734054790_s238_s1
    -rw-r----- 1 oracle dba 209723392 Jan 18 00:02 database_f734055136_s247_s1
    -rw-r----- 1 oracle dba 73408512 Jan 18 00:02 database_f734055143_s248_s1
    -rw-r----- 1 oracle dba 67117056 Jan 18 00:02 database_f734055146_s249_s1
    -rw-r----- 1 oracle dba 4194312192 Jan 18 00:02 database_f734054855_s239_s1
    -rw-r----- 1 oracle dba 2147491840 Jan 18 00:02 database_f734054975_s241_s1
    -rw-r----- 1 oracle dba 3221233664 Jan 18 00:02 database_f734054920_s240_s1
    drwxr-xr-x 2 oracle dba 4096 Jan 18 00:00 logs
    -rw-r----- 1 oracle dba 18710528 Jan 17 00:15 controlfile_c-1911789030-20110117-00
    -rw-r----- 1 oracle dba 1343488 Jan 17 00:15 database_f740621746_s624_s1
    -rw-r----- 1 oracle dba 2958848 Jan 17 00:15 database_f740621745_s623_s1
    -rw-r----- 1 oracle dba 6415990784 Jan 17 00:15 database_f740620829_s622_s1
    -rw-r----- 1 oracle dba 172391424 Jan 17 00:00 database_f740620814_s621_s1

    george3 wrote:
    Ok, perhaps its my understanding of RMAN that is at fault. From the logs :
    Starting recover at 18-JAN-11
    channel m1: starting incremental datafile backup set restore
    channel m1: specifying datafile copies to recover
    recovering datafile copy file number=00001
    name=/exlibris1/rmanbackup/database_f734055020_s242_s1
    recovering datafile copy file number=00002
    name=/exlibris1/rmanbackup/database_f734055045_s243_s1
    it seems to make backup copies of the datafiles every night, so the creation of these large files is normal ?Above results indicate that you have full (incremental level 0) backup(all datafiles copies ) and there happen update/recover (applying incremental level 1) backup.So there was happen applying */exlibris1/rmanbackup/database_f734055045_s243_s1* inremental backup to full(level 1) backup.And size should be normal
    Why is it making copies of the datafiles even on days of incrementals ?
    Because after getting level 1 backup and need applying and every day will apply one incremental backup.

  • Large number of JSP performance [repost for grandemange]

    Hi,
    a colleague of me made tests with a large number of JSP and identified a
    performance problem.
    I believe I found a solution to his problem. I tested it with WLS 5.1 SP2
    and SP3 and MS jview SDK 4.
    The issue was related to the duration of the initial call of the nth JSP,
    our situation as we are doing site hosting.
    The solution is able to perform around 14 initial invocations/s no matter if
    the invocation is the first one or the
    3000th one and the throughput can go up to 108 JSPs/s when the JSP are
    already loaded, the JSPs being the
    snoopservlet example copied 3000 times.
    The ratio have more interest than the values as the test machine (client and
    WLS 5.1) was a 266Mhz laptop.
    I repeat the post of Marc on 2/11/2000 as it is an old one:
    Hi all,
    I'm wondering if any of you has experienced performance issue whendeploying
    a lot of JSPs.
    I'm running Weblogic 4.51SP4 with performance pack on NT4 and Jdk1.2.2.
    I deployed over 3000 JSPs (identical but with a distinct name) on myserver.
    I took care to precompile them off-line.
    To run my tests I used a servlet selecting randomly one of them and
    redirecting the request.
    getServletContext().getRequestDispatcher(randomUrl).forward(request,response);
    The response-time slow-down dramaticaly as the number of distinct JSPs
    invoked is growing.
    (up to 100 times the initial response time).
    I made some additional tests.
    When you set the properties:
    weblogic.httpd.servlet.reloadCheckSecs=-1
    weblogic.httpd.initArgs.*.jsp=..., pageCheckSeconds=-1, ...
    Then the response-time for a new JSP seems linked to a "capacity increase
    process" and depends on the number of previously activated JSPs. If you
    invoke a previously loaded page the server answers really fast with no
    delay.
    If you set previous properties to any other value (0 for example) the
    response-time remains bad even when you invoke a previously loaded page.SOLUTION DESCRIPTION
    Intent
    The package described below is design to allow
    * Fast invocation even with a large number of pages (which can be the case
    with Web Hosting)
    * Dynamic update of compiled JSP
    Implementation
    The current implementation has been tested with JDK 1.1 only and works with
    MS SDK 4.0.
    It has been tested with WLS 5.1 with service packs 2 and 3.
    It should work with most application servers, as its requirements are
    limited. It requires
    a JSP to be able to invoke a class loader.
    Principle
    For a fast invocation, it does not support dynamic compilation as described
    in the JSP
    model.
    There is no automatic recognition of modifications. Instead a JSP is made
    available to
    invalidate pages which must be updated.
    We assume pages managed through this package to be declared in
    weblogic.properties as
    weblogic.httpd.register.*.ocg=ocgLoaderPkg.ocgServlet
    This definition means that, when a servlet or JSP with a .ocg extension is
    requested, it is
    forwarded to the package.
    It implies 2 things:
    * Regular JSP handling and package based handling can coexist in the same
    Application Server
    instance.
    * It is possible to extend the implementation to support many extensions
    with as many
    package instances.
    The package (ocgLoaderPkg) contains 2 classes:
    * ocgServlet, a servlet instantiating JSP objects using a class loader.
    * ocgLoader, the class loader itself.
    A single class loader object is created.
    Both the JSP instances and classes are cached in hashtables.
    The invalidation JSP is named jspUpdate.jsp.
    To invalidate an JSP, it has simply to remove object and cache entries from
    the caches.
    ocgServlet
    * Lazily creates the class loader.
    * Retrieves the target JSP instance from the cache, if possible.
    * Otherwise it uses the class loader to retrieve the target JSP class,
    create a target JSP
    instance and stores it in the cache.
    * Forwards the request to the target JSP instance.
    ocgLoader
    * If the requested class has not the extension ocgServlet is configured to
    process, it
    behaves as a regular class loader and forwards the request to the parent
    or system class
    loader.
    * Otherwise, it retrieves the class from the cache, if possible.
    * Otherwise, it loads the class.
    Do you thing it is a good solution?
    I believe that solution is faster than standard WLS one, because it is a
    very small piece of code but too because:
    - my class loader is deterministic, if the file has the right extension I
    don't call the classloader hierarchy first
    - I don't try supporting jars. It has been one of the hardest design
    decision. We definitely need a way to
    update a specific page but at the same time someone post us NT could have
    problems handling
    3000 files in the same directory (it seems he was wrong).
    - I don't try to check if a class has been updated. I have to ask for
    refresh using a JSP now but it could be an EJB.
    - I don't try to check if a source has been updated.
    - As I know the number of JSP I can set pretty accurately the initial
    capacity of the hashtables I use as caches. I
    avoid rehash.
    Cheers - Wei

    I dont know the upper limit, but I think 80 is too much. I have never used more than 15-20. For Nav attributes, a seperate tables are created which causes the Performance issue as result in new join at query run time. Just ask your business guy, if these can be reduced.One way could be to model these attributes as seperate characteristics. It will certainly help.
    Thanks...
    Shambhu

  • Number range interval for characteristic 0CSTELMNTYP is missing

    All,
       I'm moving the infoobject 0COSTELMNT, which has as a nav attribute 0CSTELMNTYP.  The transport moved fine to my Q environment, but not to prod,  The error was as above and
    Number range interval for characteristic 0CSTELMNTYP is missing
    Message no. BRAIN049
    Diagnosis
    The interval for number range object BIM0001194 of characteristic 0CSTELMNTYP is missing. The interval is created automatically, when you run the system in the BW clients.
    Procedure
    Execute the corresponding action again in the BW clients, so that the interval is created automatically. Then you can drag numbers for the characteristic 0CSTELMNTYP to other clients.
    I don't understand this - what do I need to do?
    Thanks.
    Dave

    Hi Kedar,
    I got the same issue which you mentioned and i ran the program to activate the master data object and it went fine in Quality system. But the problem here is iam unable to see the attribute data in the master data info object.
    For example i added ZCNTYCODE nav. attribute to 0VENDOR and moved to quality and i got the activation problem issue for 0VENDOR and i ran the program which you mentioned in Quality and the activation went fine.
    When i loaded the data for 0VENDOR iam not able to see the ZCNTYCODE data in Vendor master data P table. And iam able to see the master data in the table which we get from source LFA1 and in RSA3 and also in the PSA. Can you or anyone faced this type of issue.
    Thanks for you help in Advance.
    Sudhakar

  • Number range missing for characteristic 0AF_SEAN after importing request

    Hii friends,
    I m facing one problem while transporting request from development to quality , the request contains info object 0MATERIAL in which i have added new attribute 0AF_SEAN.
    Then error which i m getting is ,
    Number range interval for characteristic 0AF_SEAN is missing
    Error when activating InfoObject 0MATERIAL
    Characteristic 0MATERIAL: Error when generating master data routines
    Error when resetting InfoObject 0MATERIAL to the active version
    I am waiting for ur positive reply.
    Thanks & Regds,
    Kavan Shah

    hi Kavan,
    try
    transaction rsrv ? >> all elementry tests >> master data >> compare number range and maximum sid >> execute. press button 'correct error' if any error exists
    Error calling number range object 0MATERIAL for dimension D4 ( )
    or check oss note 903291 ?
    Manual solution
    Get the name of the number range object from the error message shown above, for example, RS_OBJNUMG. Start Transaction SNRO. To display the intervals, use the number range function key and enter the name of the number range object as the object name. Change the intervals. The following from-numbers and to-numbers apply:
    - RS_OBJNUMG: 0000001 - 9999000
    - RSDDAGCUBE: 100000 - 199999
    - RSSBRAUTH: 00000001 - 99999999
    BW 3.0B
    Import Support Package 30 for 3.0B
    BW 3.10 Content
    Import Support Package 24
    BW 3.50
    Import Support Package 16
    BW 7.0
    Import Support Package 07
    Transporting DataSources
    hope this helps.

  • How to handle a large number of query parameters for a Browse screen

    I need to implement an advanced search functionality in a browse screen for a large table.  The table has 80+ columns and therefore will have a large number of possible query parameters.  The screen will be built on a modeled query with all
    of the parameters marked as optional.  Given the large number of parameters, I am thinking that it would be better to use a separate screen to receive the parameter input from the user, rather than a Popup.  Is it possible for example to have a search
    button on the browse screen (screen a) open a new screen (screen b) that contains all of the search parameters, have the user enter the parameters they want, then click a button to send all of the parameters back to screen a where the query is executed and
    the search results are returned to the table control?  This would effectively make screen b an advanced modal window for screen a.  In addition, if the user were to execute the query, then want to change a parameter, they would need to be able to
    re-open screen b and have all of their original parameters still set.  How would you implement this, or otherwise deal with a large number of optional query parameters in the html client?  My initial thinking is to store all of the parameters in
    an object and use beforeShown/afterClosed to pass them between the screens, but I'm not quite sure how to make that work.  TIA

    Wow Josh, thanks.  I have a lot of reading to do.  What I ultimately plan to do with this (my other posts relate to this too), is have a separate screen for advanced filtering that also allows the user to save their queries if desired. 
    There is an excellent way to get at all of the query information in the Query_Executed() method.  I just put an extra Boolean parameter in the query called "SaveQuery" and when true, the Query_Executed event triggers an entry into a table with
    the query name, user name, and parameter value pairs that the user entered.  Upon revisiting the screen, I want the user to be able to select from their saved queries and load all the screen parameters (screen properties) from their selected query. 
    I almost have it working.  It may be as easy as marking all of the screen properties that are query parameters as screen parameters (not required), then passing them in from the saved query data (filtered by username, queryname, and selected
    item).  I'll post an update once I get it.  Probably will have some more questions as I go through it.  Thanks again! 

  • How to remove large number of "sent" but still keep them available in case I need to search for something?

    Over several years, I have accumulated some 1400 messages in the SENT folder.
    Lately my Thunderbird is taking a long time for some actions that used to be instant, like selecting another message to read or delet.
    I regularly remove incoming messages to keep the in-box under 300, and I have almost nothing in the other folders.
    I tried changing the "local folders" to a different directory, but when I do that, all the old "sent" messages become un-available for searching and potentially continuing a conversation.
    Is there a way to move the large number of "sent" messages from the "local folders" but still keep them available for access from within Thunderbird, in case I need something from past chains of replies?

    https://wiki.mozilla.org/Thunderbird:Testing:Antivirus_Related_Performance_Issues#McAfee

  • Best practice for handling data for a large number of indicators

    I'm looking for suggestions or recommendations for how to best handle a UI with a "large" number of indicators. By large I mean enough to make the block diagram quite large and ugly after the data processing for each indicator is added. The data must be "unpacked" and then decoded, e.g., booleans, offset binary bit fields, etc. The indicators are updated once/sec. I am leanding towards a method that worked well for me previously, that is, binding network shared variables to each indicator, then using several sub-vis to process the particular piece of data and write to the appropriate variables.
    I was curious what others have done in similar circumstances.
    Bill
    “A child of five could understand this. Send someone to fetch a child of five.”
    ― Groucho Marx
    Solved!
    Go to Solution.

    I can certainly feel your pain.
    Note that's really what is going on in that png  You can see the Action Engine responsible for updating the display to the far right. 
    In my own defence: the FP concept was presented to the client's customer before they had a person familliar with LabVIEW identified.  So I worked it this way from no choice of mine.  I knew it would get ugly before I walked in the door and chose to meet the challenge head on anyway.  Defer Panel Updates was my very good friend.  The sensors these objects represent were constrained to pass info via a single ZigBee network so I had the benefit of fairly low data rates as well but even changing view (Yes there is a display mode that swaps what information is displayed for each sensor) was updated fast enough that the user still got a responsive GUI.
    (the GUI did scale poorly though!  That is a lot of wires!  I was greateful to Jack for the Idea to make align and distribute work on wires)
    Jeff

  • Possible bug for emails with a large number of recipients

    When receiving emails which was also sent to a large number of people (more than can fit in two lines), it then displays the text 'and 10 more..." or however more there are. Clicking on this text is supposed to then display all the email addresses the email was sent to. However clicking on this is for some reason doing nothing. Is this a bug, or am I the only one having this problem?

    When receiving emails which was also sent to a large number of people (more than can fit in two lines), it then displays the text 'and 10 more..." or however more there are. Clicking on this text is supposed to then display all the email addresses the email was sent to. However clicking on this is for some reason doing nothing. Is this a bug, or am I the only one having this problem?

  • Status and messaging for systems with a large number of classes

    Very common dilemma we coders face when creating
    systems involving a large number of classes:
    Any standard framework to take care of the global status of the whole application and of gui subsystems
    for both status handling and report messages to users?
    Something light if possible...and not too much threads.

    Ah, I see,
    I found JPanel with CardLayout or a JTabbedPane very good for control of several GUI in an application - as alternative organization tool I use a JTree, which is used for both, selecting and organizing certain tasks or data in the application - tasks are normally done with data associated with them (that is, what an object is for), so basically a click onto a node in this JTree invokes an interface method of that object (the userObject), which is associated with this node.
    Event handling should be done by the event-handling-thread only as far as possible - it is responsible for it so leave this job to it. This will give you control over the order in which the events are handled. Sometimes it needs a bit more work to obey this rule - for example communication coming from the outside (think of a chat channel for example) must be converted to an event source driven by a thread normally. As soon as it is an event source, you can leave it's event handling to the event handling thread again and problems with concurrent programming are minimized again.
    It is the same with manipulating components or models of components - leave it to the event handling thread using a Runnable and SwingUtilities.invokeLater(Runnable). This way you can be sure that each manipulation is done after the other in the order you have transferred it to the event handling thread.
    When you do this consequently most of your threads will idle most of the time - so give them a break using Thread.sleep(...) - not all platforms provide preemptive multitasking and this way it is garanteed, that the event handling thread will get a chance to run most of the time - which results in fast GUI update and fast event handling.
    Another thing is, that you should use "divide and conquer" also within a single GUI panel - place components in subpanels and transfer the responsibility for the components in this panel to exactly these subpanels - think of a team manager which makes his employees work together. He reports up to his super manager and transfers global order from his boss into specific tasks by delegation to the components, he is managing. If you have this in mind, when you design classes, you will have less problem - each class is responsible for a certain task - define it clearly and define to whom it is reporting (its listeners) and what these listeners may be interested in.
    When you design the communication structure within your hierarchy of classes (directors, managers, team managers and workers) have in mind, that the communication structure should not break the management hierarchy. A director gives global orders to a manager, which delegates several tasks to the team managers, which make their workers do what is needed. This structure makes a big company controlable by directors - the same principles can also keep control within an application.
    greetings Marsian

  • Which is the best way for posting a large number of records?

    I have around 12000 register to commit to dababase.
    Which is the best way for doing it?
    What depends on ?
    Nowadays I can't commit such a large number of register..The dabatase seems hanged!!!
    Thanks in advance

    Xavi wrote:
    Nowadays I can't commit such a large number of registerIt should be possible to insert tens of thousands of rows in a few seconds using an insert statement even with a complex query such as the all_objects view, and commit at the end.
    SQL> create table t as select * from all_objects where 0 = 1;
    Table created.
    Elapsed: 00:00:00.03
    SQL> insert into t select * from all_objects;
    32151 rows created.
    Elapsed: 00:00:09.01
    SQL> commit;
    Commit complete.
    Elapsed: 00:00:00.00
    I meant RECORDS instead of REGISTERS.Maybe that is where you are going wrong, records are for putting on turntables.

Maybe you are looking for