Validation logic for US-ASCII characters

Hi folks!
I hope this question would be acceptable for all great forum members.
I just would like to know if there is a better way to validate the code point range for US-ASCII character set or
my validation logic is wrong for some reason.
Please take a look at the following validation method sample code.
By using this validation method,
I'd like a user passing a value of any other character set to be rejected for further execution.
My idea is the code point range for the US-ASCII character set is from 0 to 127.
So each character value of each String value is compared to
the maximum code point for the US-ASCII character set.
  private void validateEachElementValueHasOnlyUSASCIIChars(
      String[] aFieldValuesArray) {
    for(String fieldValue:aFieldValuesArray){
      char[] charArray = fieldValue.toCharArray();
      for(char c:charArray){
        if(c > 0x007F){// US-ASCII character set's maximum code point.
          throw new IllegalArgumentException(
              "the value you passed" +
              "("+Arrays.toString(charArray)+") contains a " +
              "character of the invalid character set.");
  }Similar topic: [http://forums.sun.com/thread.jspa?forumID=31&threadID=751191]
Thanks for your time.
Kenjiro
Edited by: kenjiro on Sep 26, 2008 5:49 AM

Yep, looks ok to me. Depending on your requirement you might also disallow some of the non-printing control characters that are in ASCII (all of which are below 32, IIRC).

Similar Messages

  • Implementing Custom Validation Logic for Payment Disbursements

    Hi All,
    Is there anyone here, who has implemented Custom Validation Logic for disbursements in Oracle payments.
    Theory: Under Payments Setup Administrator -> Formats -> Formats, we can setup a payment format and even attach a XML template. Apart from this we can setup Pre-defined validations or user Defined validations.
    For Pre-defined validations / seeded validations, Oracle has provided an API
    IBY_VALIDATIONSETS_PUB.evaluateCondition. We can use this API, and call a custom package by using the keyword 'CUSTOM' in the call.
    Has anyone played around with this package and successfully implemented custom validations. Does anyone have a sample code on how to implement this. The Payment Manager Implementation guide is of little help.
    Thanks

    This document was very helpful to me in creating a custom validation.
    http://docs.oracle.com/cd/E18727_01/doc.121/e13416/T387353T419293.htm

  • Can you please help me with validation logic for Events in Table maintenance generator

    Can you please help me with validation logic for Events in Table maintenance generator,i.e if i enter record in 1st internal table then automatically 2nd internal table should be updated.

    Hi Glen Anthony,
        Thanks for replay,
         I used foreign key relationship between those 2 internal tables....
    I used event 05: When creating a new entry. I want to know the custom logic by which my 2nd Internal table gets automatically updated when i update my 1st Internal table
    Thanks Glen.

  • Validation for non-ASCII characters

    Hi all,
    Requirement: I have to apply a validation on on fields like Name and Address in applicationdefination.xml. When a user types non-ASCII characters and navigates to next page then it should display the error message. Thus, I have to restrict my user to ASCII values only.
    Present Situation: I'm using regular expression for this problem. In Jheadstart there is an option regular expression under the heading Validation. I have written following values in regular expression and Regular Expression Error Message options.
    Regular Expression
    ^\s*[\w\.\,\-\_\(\)\#\'\/\\\ u0022\u0026\*\;\:\s]+\s*$
    Regular Expression Error Message
    It is important to note that foreign characters are not accepted on our system. Please ensure only standard English letters are entered
    Since, i was getting error in jspx page due to double quotes(") and ampercent(&), So i have replaced the double quotes(") and amprecent(&) by their unicodes. Thus, the expression has become like ^\s*[\w\.\,\-\_\(\)\#\'\/\\\u0022\u0026\*\;\:\s]+\s*$.
    This expression is validating many characters like Ã,µ,Ç,Ï,Ö,§,¥,{,} but not all non ASCII characters like ѓ є ѕ ї Њ Щ Ώ Ω Ϊ Ά Ή Θ Λ Ξ Π τ ẫ ờ Ỡ Ứ Ỷ ự Ẁ ỹ ị Ọ ň ũ ť ţ Έ Ϊ ﻍ. Thus, its not fulfilling the requirement.
    Please suggest some valid solution to this problem. It’s very urgent.

    Hi,
    The validation seems to be performed in Java or Javascript depending on the layout (I'm sorry I can't remember the exact details). The expression suggested above by theEternalStudent works very well in Java, but not in Javascript.
    We came up with an expression which works in both. It rejects strings which contain &# by doing a lookahead before the main pattern - you might want to expand this to look for &#nnn; but for our purposes &# is enough.
    Here is the "platform neutral" solution:
    (?!.*\u0026#.*)^[\w\.\,\-\_\(\)\#\'\/\\\u0022\u0026\*\;\:\s]+$
    I think in future we will write a javascript function and amend the templates to call it directly.
    thanks,
    Michael

  • Contains query fails for extended ascii characters

    I have an Oracle 9.2 instance whose characterset is WE8MSWIN1252. I'm using the same characterset on my client. If I have a LONG column that contains extended-ascii characters (the example I'm using has the Euro character '€', but I've seen the same problem with other characters), and I'm using the Intermedia service to index that column, then this select statement returns no records even though it should find several:
    select id from table1 where (contains(long_col,'€',1) > 0);
    However, the same select statement looking for something else, like 'e', works just fine.
    What am I doing wrong? I can do a "like" query against a VARCHAR2 column with a Euro character, and it works correctly. I can do a "dbms_lob.instr" query against a CLOB column with a Euro character, and it also works. It's just the "contains" query against a LONG column that fails.

    There are a number of limitations in using Long datatypes. If you check the SQL Reference you will see: "Oracle Corporation strongly recommends that you convert LONG columns to LOB columns as soon as possible. Creation of new LONG columns is scheduled for desupport.
    LOB columns are subject to far fewer restrictions than LONG columns. Further, LOB functionality is enhanced in every release, whereas LONG functionality has been static for several releases."

  • Preview search doesn't work for all ASCII characters

    The title pretty much says it all. The behavior is completely broken. Any searches for words that contain a non-letter, non-number ASCII, non-standard punctuation character, e.g., @#$%^&*()+, kind of ignores the special characters. For example:
    Find: C++
    Finds all pages with the text "C++", but doesn't highlight that text. Instead, it highlights all "C" and "c" characters, and the find counts for those pages are the number of C characters. Note that it neither highlights nor counts C characters on pages that don't contain "C++"
    Find: "C++"
    Finds about half the pages and the find count is correct, but it highlights only C characters that are followed by ++ OR other special characters, e.g. it highlights the C in "C#".

    mentalwasteland wrote:
    I simply want it to work as the documentation says it works.
    By all means tell Apple it does not perform according to their documentation when you ask them to fix it.
    Does Adobe Reader do any better?  I don't know if it might have its own search system....

  • How to create a native KeyEvent for non ASCII characters

    Hello
    i need to create a native KeyEvent with my application. I know that it is possible to send such a event with the Robot-Class. But how do i send an event for a none ASCII character such as the german � (O with Diaeresis)?
    I also know that a combination of KeyEvent.VK_DEAD_DIAERESIS + KeyEvent.VK_O will get me the desired result, but i want just to pass the charachter � and let Java create the correct KeyCode(s).
    Does anyone knows a solution?
    BTW: AWTKeyStroke always returns 0 for the �
    Thanks
    Matt

    Hi, James.
    Unfortunately, none of the F-keys can be set to the functions of the Volume keys using the Keyboard Shortcuts preference pane.
    Mac OS X does not include native support for assigning a macro or script to either an Fkey or other keyboard shortcut. For that, you need a third-party tool, like iKey, Quickeys, or Spark, the latter suggested by DPSG-Scout.
    You'd also need a set of scripts to assign to the keys, e.g. scripts to increase, decrease, or mute the volume. Those are tricky to write since System Preferences does not expose much to AppleScript. One would need to use the "GUI Scripting (System Events)" technique to create a script to change the settings in the Sound preferences pane of System Preferences.
    Automator does offer a "Set Computer Volume" action, which allows you to adjust the various Volume settings (Output volume, Alert Volume, Input Volume) to a specific set of levels, including Mute for Output Volume. However, it won't handle the idea of pressing the same key multiple times to either increase or decrease the volume, ala the Volume Adjustment keys. In your case, it's most useful for Mute. Automator Workflows can be saved as applications, which I believe can then be assigned to Fkeys in iKey or QuickKeys.
    Someone may have written a more advanced Automator action for adjusting volume than that provided with Mac OS X. Search Apple's Automator Action Downloads. If you find one, instructions on adding actions to Automator can be found in Automator Help.
    You could search ScriptBuilders to see if someone has already written scripts for the functions you need, then use iKey or Quickeys to assign such to keyboard shortcuts.
    Finally, you could solve the problem another way: buy a set of external speakers that has its own separate volume and mute control. For example, my trusty old Monsoon MM1000s have their own volume control with a mute button.
    Good luck!
    Dr. Smoke
    Author: Troubleshooting Mac® OS X

  • How to query for extended ASCII characters in column value

    Hi All
    Sorry if this has been answered before. I tried searching but none of them seems to work for me.
    I am trying to search for inverted ? in my column.
    I am using the following query
    select *
    from table_name
    where regexp_like (description , '¿' )

    Did you try:
    like '%¿%'
    create table table_name (description varchar2(100))
    insert into table_name values ('jh¿¿gagd')
    insert into table_name values ('jhga345gd')
    insert into table_name values ('j1231232hgagd¿')
    select *
    from table_name
    where regexp_like (description , '¿' )
    DESCRIPTION
    jh¿¿gagd
    j1231232hgagd¿
    select * from table_name
    where description like '%¿%'
    DESCRIPTION
    jh¿¿gagd
    j1231232hgagd¿Edited by: user130038 on Sep 8, 2011 12:29 PM

  • What is current CommSuite support for non-ASCII passwords?

    Hello all,
    Some of our users managed to change their passwords to non-ASCII strings (via replication from MSAD by ISW) and no longer have access to their communications services.
    While replicating the problem, I have set a (UTF-8 non-ASCII) string as my password in DSEE directly, and *can* log in to Convergence with this password. However, if I change the working password to a non-ASCII string from Convergence itself - it is accepted during the secondary password check, there is no error returned, SOME password is apparently saved into the LDAP directory, but neither of the original non-ASCII plaintext strings can be used for login back into Convergence. Restoration of access is only doable by admin at this point.
    Checking email by IMAP from Thunderbird no longer works with a changed non-ASCII password (including the state when it still works for Convergence).
    Delegated Admin has an explicit check for non-ASCII characters in the password and refuses to set a misbehaving one.
    I see that among the standards supported by CommSuite, there is IMAP4rev1, and RFC 5255 refers to it as the reason that non-ASCII passwords and usernames are for now not supported, though this is expected to be a temporary state of things, and software can prepare for the future by implementing checks for valid UTF-8 strings as well.
    https://wikis.oracle.com/display/CommSuite/Messaging+Server+Supported+Standards
    http://tools.ietf.org/html/rfc5255
    5.1.  Unicode Userids and Passwords
       IMAP4rev1 currently restricts the userid and password fields of the
       LOGIN command to US-ASCII.  The "userid" and "password" fields of the
       IMAP LOGIN command are restricted to US-ASCII only until a future
       standards track RFC states otherwise.  Servers are encouraged to
       validate both fields to make sure they conform to the formal syntax
       of UTF-8 and to reject the LOGIN command if that syntax is violated.
       Servers MAY reject the LOGIN command if either the "userid" or
       "password" field contains an octet with the highest bit set.
       When AUTHENTICATE is used, some servers may support userids and
       passwords in Unicode [RFC3490] since SASL (see [RFC4422]) allows
       that.  However, such userids cannot be used as part of email
       addresses.
    So, the main question at this point is: does or does not all of the CommSuite stack support non-ASCII passwords?
    If no - please confirm, so we can instruct the users to not create problems for themselves (and maybe manage to set up some policy to not accept non-ASCII passwords to MSAD/DSEE in the first place).
    If yes - what should be done to enable support in Convergence/IMAP/SMTP/XMPP/WCAP/WABP/... services - perhaps, setting the LANG/LC_ALL locale environment variables or equivalent JVM flags for UTF-8 in server startup scripts, etc.? (I know that DSEE ldapsearch requires either envvars or a command-line flag for charset encoding of values, so I figure similar quirks may be relevant for some other software)
    Thanks in advance for either response,
    //Jim Klimov

    I can't respond for the suite, but the Messaging Server product should work with UTF-8 usernames and passwords as long as the standard SASL authentication mechanisms that are documented to use UTF-8 are used (e.g. SASL PLAIN). IMAP LOGIN may work fine with UTF-8 as well even though that's non-standard. We do not implement SASLprep, however, so the strings provided by the client to the server must be identical UTF-8 strings for authentication to succeed. If they are provided in a different decomposition, different canonical form or non-standard charset, that's not supported and will fail. We don't test this scenario extensively, so you may encounter bugs (that we'd have to prioritize and fix as with other bugs). Messaging Server recently implemented a restricted option (broken_client_login_charset) for a customer who was stuck with broken clients that sent ISO-8859-1 for the IMAP login command arguments.

  • Where to put validation logic...

    Hi,
    We have a lot of validation logic in our database. Next to the normal foreign key, check, unique constraints, we use triggers for validation (which throw errors with custom error codes when needed).
    On top of the database we are building a fairly large JClient based application using the BC4J framework. BC4J offers the possibility to use validation beans for validating the entity row. We are wondering if we should use this. If we implement the validation logic for each entity we are actually duplicating the validation logic that's already available in the database. And some database constraints are easily implemented using the standard validation beans (like not null and unique checks), but more complex validation rules can sometimes only be done in the database (so we need to use the database for validation using a method validation bean).
    Like I said this means a lot of duplicating validation logic. Which first of all is an inefficient thing to do. Second validation logic can change over time, which means we need to change the logic on two places. It's easy to forget to change the logic on one place.
    We found a way to somewhat use the database validation logic for validating entity rows. We just simple insert a row in the database and post the changes. If the new row contains an error an exception is thrown. This JBO exception contains a reference to the SQL exception. Which can be used to retrieve the database error code. Unfortunately we have not found a good way (yet) to find out where the error occured (in which trigger for example). Error codes don't have to be unique so it's important to know in which context they occured. Another problem with this method is we can't do an early validate on a limitted set of row data. E.g. we sometimes use wizards in our application for adding new rows. We really want to check the data after each step in the wizard (step by step). This is ofcourse impossible with the database logic. This means we again need to create the same validation logic again. I don't think it's even possible to use the row.validate() method for this (and ofcourse implement the validation logic in the entity too), because this method validates the complete row.
    In short, where is the best way to put validation logic? And (how) can this be done without duplicating (to much) logic?
    Regards,
    Peter
    P.S.
    Does anyone know how to retrieve the context of an SQL exception?

    Do you users have access to the data/tables outside
    of using your application? (as in using sqlplus?)Not at the moment.
    Ideally the database should just be going
    constraint--data integrity type of checks. The
    business rules would be in the business layer not
    persitance layer..So in this case, they would be in
    the bc4j entity objects.. Well most of them are simple integrity constraints, but sometimes we have to check things using triggers which are way more complicated and need extensive database access. It's more logical to do that in the database. If the trigger runs into a problem it raises an application error. We are hoping we can use this error somehow in our application.
    That assumes that the only way a user can modify that
    data is through your application only.Well it is, and if another application is written for it the BC4J entities etc. will probably be re-used, but like I said there are other reasons for not having all the business-rules in the BC4J layer.
    by the way, are you migrating an existing application
    that already has all the business rules in the
    database?No, it's a brand new application, but we started out with defining a lot of business rules in the database because that seems a more logical place to put them (to us). I understand that some business-rules can better be implemented in the BC4J entities, but it's not that easy for every business rule we have.
    Regards,
    Peter

  • Given filename or path contains Unicode or double-byte characters.Retry using ASCII characters for filename and path What does this mean? it happen when I publish an OAM

    Given file name or path contains Unicode or double-byte characters. Retry using ASCII characters for filename and path
    What does this mean? It is happening when I try to publish an OAM for Dreamweaver.
    Also: How can I specify the browser in Edge Animate? It is just going wherever. Are there no Preferences for Edge Animate?
    BTW. Just call it Edge. Seriously. Do you call it Illustrator Draw? Photoshop Retouching?

    No, my file name is mainContent.oam
    My project name is mainContent.an
    This error happens when I try to import into Dreamweaver. Sorry, I wasn't clear on that earlier.
    I thought maybe it was because I had saved my image as a png. So re-saved as a svg, still get the error.
    DO I have a setting is Dreamweaver CC that is wrong? Should I try this in Dreamweaver CS6? I might try that next.
    Why is this program so difficult? I know Flash. I know After Effects. I can work the timeline part just great. It's always in the export that I have problems.
    On a MacPro, 10.7.
    Are you an Adobe person or just a nice helper?

  • Search for users and non-ASCII characters

    I am having a little issue with the "Accounts - Find Users" functionality. The search breaks on what I assume is non-ASCII characters (we use the following three up here in Denmark: �, �, �). To be precise, I have a user with the first name "J�rgen". Searching for first names starting with "J" works just fine but "J�" returns zero matches.
    My setup is with two machines, one (A) holding the MySQL database and one (B) serving Identity Manager on top of tomcat.
    Both A and B are RHEL boxes, and both have da_DK.UTF-8 as default locale.
    MySQL's /etc/my.cnf file has the following entry (as recommended in create_waveset_tables.mysql):
    [mysqld]
    default-character-set=utf8
    default-collation=binFor clarity, some functionality works just fine in Identity Manager with these non-ASCII characters such as adding a user whose name contains non-ASCII characters (not only ��� but also � for example). At the moment, it appears to be the search functionality which is not working correctly as I would expect it to. I'm still on the fence concerning whether I've missed something in terms of configuration, or whether this is a limitation.
    Does anyone know whether this problem is on my side or the software's side?

    I am having a little issue with the "Accounts - Find Users" functionality. The search breaks on what I assume is non-ASCII characters (we use the following three up here in Denmark: �, �, �). To be precise, I have a user with the first name "J�rgen". Searching for first names starting with "J" works just fine but "J�" returns zero matches.
    My setup is with two machines, one (A) holding the MySQL database and one (B) serving Identity Manager on top of tomcat.
    Both A and B are RHEL boxes, and both have da_DK.UTF-8 as default locale.
    MySQL's /etc/my.cnf file has the following entry (as recommended in create_waveset_tables.mysql):
    [mysqld]
    default-character-set=utf8
    default-collation=binFor clarity, some functionality works just fine in Identity Manager with these non-ASCII characters such as adding a user whose name contains non-ASCII characters (not only ��� but also � for example). At the moment, it appears to be the search functionality which is not working correctly as I would expect it to. I'm still on the fence concerning whether I've missed something in terms of configuration, or whether this is a limitation.
    Does anyone know whether this problem is on my side or the software's side?

  • UTF8 Byte Size for ASCII Characters

    In a UTF8 database, are ASCII characters stored as a single byte, therefore using the minimum amount of db space possible? Or are they stored at a fixed amount of 2 or 3 bytes, wasting db space unnecessarily?

    Hi,
    The Ascii characters are single bytes. So don't worry about space.

  • GSSName  is corrupted for non ascii chars

    Hi,
    I have a setup where a web application is deployed to use SPNEGO for user authentication ( using kerberos V ) and authorization.
    We have several users with non english characters in the user ID and even though kerberos authentication succeeds for such users ( KDC / Active Directory is returning valid kerberos ticket which the client embeds in the SPNEGO token). Hoowever, on passing the SPNEGO token to GSS API and extracting the user name from GSS API returns incorrect user name. All non ascii characters in the user name are replaced with some junk byte sequences.
    We use JGSS API (with JRE 1.4.08) for extracting the SPNEGO token and create a GSS secruity context object. Later, the GSS Name is extracted from the GSS context object.
    Currently I am tesitng the SPNEGo authentication for a user with user ID 123<sp char> . The <sp char> 's unicode value is FE and UTF-8 encoded byte sequence is C3 BE. However, if I invoke 'export' method of the GSSName object and examine the returned byte sequence, instead of C3 BE, the byte sequence EF BF BD EF BF BD is present. The byte sequence for other english characters are proper.
    Is this a defect in GSS-API ? Or am I not using GSS properly?
    Do I need to have any special setup / configuration for using JGSS with kerberos V for users with non ascii characters in the user ID?
    Please advise.
    Regards,
    Jayaram.
    Message was edited by:
    s_jayaram_s

    I understand that this is an older thread.
    We have spent lot of time on the internet to find out any possible workarounds or permanent solutions to enable Multi-Byte character support for username / password / SPN with Kerberos and Java . But no luck so far :(
    Are there any new updates on this i18n issue ?
    Thanks,
    Venkatesh

  • Logic for correcting strings with different barckets

    can any one tell me the simple logic for correcting strings with different barckets
    suppose say for eg.
    (<{abc}>defgh) should evaluate to TRUE
    (<{abc>} should evaluate to FALSE
    Things to check is
    1. number of opening brace and closing brace
    2. Their position are correct or not
    If the above two conditions are met than it is TRUE else FALSE
    i guess string function with decode and lenght function or case should do but bit difficult to develop this logic in with sql
    Thanks,
    AAK

    Hi,
    Here is non REGEXP_REPLACE version ( Unfotunately dont have 10G to feel the power of REGEXP_REPLACE.
    SQL> ed
    Wrote file afiedt.buf
      1  SELECT DECODE(SUM(CASE
      2                    WHEN Left = '(' AND Right =')' THEN 0+Count_Bracket
      3                    WHEN Left = '<' AND Right ='>' THEN 0+Count_Bracket
      4                    WHEN Left = '{' AND Right ='}' THEN 0+Count_Bracket
      5                    WHEN Left = '[' AND Right =']' THEN 0+Count_Bracket
      6                    ELSE
      7                       1
      8                    END
      9                    )
    10         ,0,'Valid','Invalid')
    11  FROM
    12  (
    13     SELECT SUBSTR(str,(len+1 - level),1) Left, SUBSTR(str,(len + level),1) Right, MOD(LENGTH(str),2)
    14     FROM
    15     (
    16      SELECT
    17         REPLACE(translate('&exp','abcdefgh','        '),' ') str,
    18         LENGTH(REPLACE(translate('&exp','abcdefgh','        '),' '))/2 len
    19      FROM DUAL
    20    )
    21    CONNECT BY LEVEL <= LEN
    22* )
    SQL> /
    Enter value for exp: (<{abc>} )
    old  17:        REPLACE(translate('&exp','abcdefgh','        '),' ') str,
    new  17:        REPLACE(translate('(<{abc>} )','abcdefgh','        '),' ') str,
    Enter value for exp: (<{abc>} )
    old  18:        LENGTH(REPLACE(translate('&exp','abcdefgh','        '),' '))/2 len
    new  18:        LENGTH(REPLACE(translate('(<{abc>} )','abcdefgh','        '),' '))/2 len
    DECODE(
    Invalid
    SQL>
    SQL> /
    Enter value for exp: <{[()]}>
    old  17:        REPLACE(translate('&exp','abcdefgh','        '),' ') str,
    new  17:        REPLACE(translate('<{[()]}>','abcdefgh','        '),' ') str,
    Enter value for exp: <{[()]}>
    old  18:        LENGTH(REPLACE(translate('&exp','abcdefgh','        '),' '))/2 len
    new  18:        LENGTH(REPLACE(translate('<{[()]}>','abcdefgh','        '),' '))/2 len
    DECODE(
    Valid
    SQL>
    SQL> /
    Enter value for exp: ([{<>}])
    old  17:        REPLACE(translate('&exp','abcdefgh','        '),' ') str,
    new  17:        REPLACE(translate('([{<>}])','abcdefgh','        '),' ') str,
    Enter value for exp: ([{<>}])
    old  18:        LENGTH(REPLACE(translate('&exp','abcdefgh','        '),' '))/2 len
    new  18:        LENGTH(REPLACE(translate('([{<>}])','abcdefgh','        '),' '))/2 len
    DECODE(
    Valid
    SQL>
    SQL> /
    Enter value for exp: (<{abc}>defgh)
    old  17:        REPLACE(translate('&exp','abcdefgh','        '),' ') str,
    new  17:        REPLACE(translate('(<{abc}>defgh)','abcdefgh','        '),' ') str,
    Enter value for exp: (<{abc}>defgh)
    old  18:        LENGTH(REPLACE(translate('&exp','abcdefgh','        '),' '))/2 len
    new  18:        LENGTH(REPLACE(translate('(<{abc}>defgh)','abcdefgh','        '),' '))/2 len
    DECODE(
    Valid
    SQL> Note: Only thing needs to care is " translate('<{[()]}>','abcdefgh',' ') " if string has some additional characters please include here.
    Regards

Maybe you are looking for

  • Error while loading master data.

    I am facing error while loading master data in a infoobject. The error comes when I run DTP fo rloading data from PSA to infoobject. The error wordings are : "MCK_PRDT1 : Data record 1 ('Bleach '): Version 'Bleach ' is not valid" Please help me in un

  • Table and field for the given object

    Hi all, We are developing a report , i need following table and fields. Could you please help me on this? Supplier No of approval required Approval status Approval pending with In approval since Thanks in advance , will appriciate any input Hareesha

  • I would like to download QuickTime Pro for iPad

    I would like to download QuickTime Pro for iPad. Please help.

  • Initialisation Period according to SAP Note: 487381

    Hi, SAP Gurus, I have a problem My fiscal year start from April to March. Unfortunately I open the period 11th i.e. February. Now I want to post in month of November ie…. in Period “08”. I checked in MMRV where back posting allowed. So I can post onl

  • Urgent Requirement for SAP CONSULTANTS - all modules.

    TOP MNC NOW HIRING Urgent Requirement for SAP CONSULTANTS - all modules. Experience    2-5 yrs. Location - Gurgaon, Hyderabad,Pune. Interested candidates please forward your CV ASAP!!!!!!!! !!!!!! send your resumes at [email protected] Thanks, Rajend