Data Validation issue with Standard Extraction

Hi Experts,
I have a requirement to publish a report in BI from ECC, all the details are coming from this DataSource : 2LIS_02_ITM.
This particular DataSource have the many fields which are from the following tables
EKKO
EKPO
MARA
EKBE
I tried replicating the DataSource in BI and created transformation extracted full data.The number of record the DSO (0PUR_O01) has is 27910.
I checked with the ECC tables. And the number of records they have are
EKKO -   79129
EKPO -  250502
MARA -   40950
EKBE - 1447470 
I just don't know how do I check whether BI data is correct with ECC.
Can anyone please tell me how do I reconcile the data.
Thanks

Hi,
you have to check the number of records for the datasource in RSA3 transaction.
Also, check the start routine at transformation level, as well as the selections if you have loaded.
Hope this helps,
Regards,
Rama Murthy.

Similar Messages

  • Validation issues with form (was: Validation)

    I am having validation issues with my form.  Even after I put in a name and last name it says the entries are invalid.  Could this be a hosting problem?  www.bussmanncomputers.net/Nx

    Below is mail form from my isp, which is Godaddy:
    This is the form that I am having some problems with configuring.
    <?php
    if ( !isset($_SERVER['SPI'])) {
            die();
    if (!isset($_SERVER['DOCUMENT_ROOT'])) {
            echo("CRITICAL: we seem to be running outside of the norm.\n");
            header("Location: http://".$_SERVER["HTTP_HOST"]."/");
            die("CRITICAL: Document root unavailable.\n");
    $request_method = $_SERVER["REQUEST_METHOD"];
    if($request_method == "GET") {
            $query_vars = $_GET;
    elseif ($request_method == "POST") {
            $query_vars = $_POST;
    reset($query_vars);
    function customsort($a,$b) {
            // $a is array for form vars, $b is comma seperated case sensitive field order
            // this is case sensitive -- good idea to hrc that.
            $data = array();
            if ( strstr($b,',') == FALSE )  {
                    $b = $b.",";
            $ordering = split(',',$b);
            foreach ($ordering as $orderitem) {
                    if ( ($orderitem != null) && ($orderitem != "") ) {
                            if (isset($a[$orderitem])) {
                                    $data[$orderitem] = $a[$orderitem];
            foreach ($a as $key=>$val) {
                    $data[$key] = $a[$key];
            return $data;
    function xmlentities($string) {
            return str_replace ( array('&', '"', "'", '<', '>'), array('&amp;', '&quot;', '&apos;', '&lt;', '&gt;'), $string);
    $t = date("U");
    $formhomedir = preg_replace('/.*\/home\/content/','',$_SERVER['DOCUMENT_ROOT']);
    $formhomedir = explode('/',$formhomedir);
    if (count($formhomedir) <= 4) {
            $formhome="/home/content/".$formhomedir[1]."/".$formhomedir[2]."/data/";
    else {
            $formhome="/home/content/".$formhomedir[1]."/".$formhomedir[2]."/".$formhomedir[3]."/".$f ormhomedir[4]."/data/";
    $file_order = ".default";
    $file_format = ".text";
    $file_interval = ".15m";
    $field_order = "";
    if (isset($query_vars['form_order'])) {
            if ($query_vars['form_order'] != "alpha") {
                    $field_order=$query_vars['form_order'];
                    $file_order=".custom";
                    $query_vars = customsort($query_vars,$field_order);
            else {
                    switch ($query_vars['form_order']) {
                            case "alpha":
                                    uksort($query_vars,'strnatcasecmp');
                                    $file_order=".alpha";
                            break;
                            default:
                                    $file_order=".default";
                            break;
    if (isset($query_vars['form_format'])) {
            switch ($query_vars['form_format']) {
                    case "csv":
                            $file_format = ".csv";
                    break;
                    case "html":
                            $file_format = ".html";
                    break;
                    case "xml":
                            $file_format = ".xml";
                    break;
                    case "text":
                    case "default":
                    default:
                            $file_format = ".text";
                    break;
    if (isset($query_vars['form_delivery'])) {
            switch ($query_vars['form_delivery']) {
                    case "hourly":
                            $file_interval = ".60m";
                    break;
                    case "hourly_digest":
                            $file_interval = ".60mc";
                    break;
                    case "daily":
                            $file_interval = ".24h";
                    break;
                    case "daily_digest":
                            $file_interval = ".24hc";
                    break;
                    case "digest":
                            $file_interval = ".15mc";
                    break;
                    case "default":
                    default:
                            $file_interval = ".15m";
                    break;
    $file = $formhome."form_".$t.$file_order.$file_format.$file_interval;
    $fp = fopen($file,"w");
    reset($query_vars);
    switch ($file_format) {
            case ".csv":
                    $csvkeys = "";
                    $csvvals= "";
                    $firsttime = "";
                    while (list ($key, $val) = each ($query_vars)) {
                            if ( ($key == "form_order") ||
                                    ($key == "form_format") ||
                                    ($key == "form_delivery") ||
                                    ($key == "redirect") ) {
                            else {
                                    if ($csvkeys != "") {
                                            $firsttime=",";
                                    $tmpkey=escapeshellcmd($key);
                                    $csvkeys = $csvkeys.$firsttime."'".$tmpkey."'";
                                    $tmpval=escapeshellcmd($val);
                                    $csvvals = $csvvals.$firsttime."'".$tmpval."'";
                    fputs($fp,"$csvkeys\n");
                    fputs($fp,"$csvvals\n");
            break;
            case ".html":
                    fputs($fp,"<table border=\"1\" cellspacing=\"1\" cellpadding=\"2\">\n");
            break;
            case ".xml":
                    fputs($fp,"<form>\n");
            break;
    reset($query_vars);
    while (list ($key, $val) = each ($query_vars)) {
            if ($key == "redirect") {
                    $landing_page = $val;
            if ( ($key == "form_order") ||
                    ($key == "form_format") ||
                    ($key == "form_delivery") ||
                    ($key == "redirect") ) {
            else {
                    switch ($file_format) {
                            case ".html":
                                    fputs($fp,"\t<tr>\n");
                                    fputs($fp,"\t\t<td><b>$key</b></td>\n");
                                    fputs($fp,"\t\t<td>$val</td>\n");
                                    fputs($fp,"\t</tr>\n");
                            break;
                            case ".csv":
                                    // content is already output
                            break;
                            case ".xml":
                                    fputs($fp,"\t<field>\n");
                                    fputs($fp,"\t\t<fieldname>".xmlentities($key)."</fieldname>\n");
                                    fputs($fp,"\t\t<fieldvalue>".xmlentities($val)."</fieldvalue>\n");
                                    fputs($fp,"\t</field>\n");
                            break;
                            case ".text":
                            default:
                                    fputs($fp,$key.": ".$val."\n");
                            break;
    switch ($file_format) {
            case ".html":
                    fputs($fp,"</table>\n");
            break;
            case ".xml":
                    fputs($fp,"</form>\n");
            break;
    fclose($fp);
    if ($landing_page != "") {
            header("Location: http://".$_SERVER["HTTP_HOST"]."/$landing_page");
    else {
            header("Location: http://".$_SERVER["HTTP_HOST"]."/");
    ?>

  • A data-binding issue with a combo box.

    Hi,
    Iu2019m having a data-binding issue with a combo box.  The field it is bound to is an integer.  The valid values in the combo box are 1, 2, and 3.  If I add a record when it is set to 1 or 2, the value gets stored correctly.  If I add a record when it is set to 3, it is stored as 1.  However, I can bring up the record just added on the form, change it to 3, and click u201CUpdateu201D and it is saved correctly as 3.  If I change it so that the valid values are 4, 5, and 6, it saves a 1 regardless of what is selected in the combo box.  It looks like the combo box is correctly bound for updates but not for adds.  The table is a master type UDO.  Any ideas?  Iu2019m somewhat committed to the field being an integer.  I'm using 2007A PL47.
    Thanks,
    Mike
    Edited by: Mike Angelastro on Jul 1, 2009 2:43 PM

    I tried a few things on my own.  The result is that I decided that it was not a good idea to use a combo box bound to an integer (numeric) field.  I donu2019t think the SDK can handle it.  The reason it was an integer field in the first place is that before I changed it to a combo box it was a group of two option buttons.  Option buttons use an integer (numeric) field.  This worked just fine until I added a third option; the SDK didnu2019t handle the third option correctly when adding records.  I thought that using a combo box instead would fix that.  I was wrong; the problem remained.  So I decided to use a character (alphanumeric) field instead.  This works just fine.
    So here is my advice:  Never use option buttons if they need to be bound to the database; a combo box will actually work better when bound to the database.  But use a character (alphanumeric) field.
    Edited by: Mike Angelastro on Jul 5, 2009 9:15 PM

  • Windows 8.1 Data reordering issue with Intel Adaptors

    According to Intel, there is a data reordering issue with their adaptors and probably this dumb WIDI software. This is from Intel site. they say some are fixed, "A future Windows 8 fix will address this issue for other Intel wireless adapters." I
    have one Nope, still broke. I get drops all the time. Brand new Toshiba laptop I7 16 gigs of ram and a SSD and a 2 gig Vid card. Would be nice to be able to play games but I get dropped all the time. Now would Microsoft quit hiding
    about this, and fix the darn thing. Also i'm a system admin for 13 years. I have build over 1000 PCs and servers. I know bad software. Please fix this. PLEASE. Its not going to just go away and its not just Toshiba, I have seen other companies with the
    same problem. If there is a Fix PLEASE POST IT. Or even a workaround I have tried everything.
    http://www.intel.com/support/wireless/wlan/sb/CS-034535.htm
     

    Hi,
    Have your first tried the software fix under this link for your network adapter?
    http://www.intel.com/support/wireless/wtech/proset-ws/sb/CS-034041.htm
    Please Note: The third-party product discussed here is manufactured by a company that is independent of Microsoft. We make no warranty, implied or otherwise, regarding this product's performance or reliability.
    Also, you can try to check if there is any driver update under Device manager from manufacture's website.
    Kate Li
    TechNet Community Support
    Yep didn't work. Still get drops all the time, had to run a Cat 5E cable to my laptop from my modem, because I have Atheros Gigabyte Lan adaptor. Works Great. The Wireless still drops all the time. Has Microsoft let out the patch to fix this or is it coming in
    April in the 8.1 patch that's coming. Funny thing is all for Widi, I don't even use widi, I got the software to do that from Samsung works better on my TV. Intel and Microsoft need to get this fixed. because their driving off gamers and that's the
    people that make sure they buy Microsoft so they can play games. With the wireless link dead and a great laptop worthless what's the point. Ive been in IT for 13 years building PCs  and Servers how I knew how to run a 60 FT Cat 5e line thru
    a 2 story house and terminate it. Most people don't. Fix the problem.  

  • ESB File Adapter data validation issue

    I have a really simple ESB process.
    File Adapter Read
    (XML format with XSD) --> Routing Service --> SOAP Call to BPEL Process
    Quite simply, the ESB uses the File Adapter to read an XML file which has an associated XSD. The contents are then routed to a BPEL Process.
    If the XML file is valid in syntax but contains an element which is not in the XSD, the file contents are pushed to the BPEL process which has validateXML set to true - thus the error is caught and handled by the BPEL process.
    This is what I want to happen.
    If the XML file is invalid in syntax - for example -
    <TAG>hello</WRONGTAG>
    The ESB process rejects the file and stops dead.
    I would rather the invalid file is pushed to the BPEL so that the BPEL can handle the error.
    Why?
    Because my error handling in the BPEL picks up the invalid XML format and emails the users to inform them there is a problem.
    Can someone help? Is this possible? Thanks.
    Message was edited by:
    IanBaird

    Hello again! You are now aware of the problem as there has been a lot of conversation between myself and Dave Berry [and you!] regarding this issue.
    For the rest of the readers - here is a brief summary of progress so far. Sorry but there has been so much communication that I think the best way is to direct you to various links.
    If an adapter fails in ESB, then it is possible to trap these failures and handle them as documented here:
    Dave Berry said:
    you need to look into the adapter rejection handlers which are documented on OTN ESB page in the exception handling lesson and the adapters OTN dev resources page http://www.oracle.com/technology/products/integration/adapters/dev_support.html
    I had also devised my own solution in the meantime, which involves reading the input file as opaque schema and passing this to a BPEL process which converts it to an XML document! This can then be validated and use standard BPEL error handling to find any problems in the file format etc.
    See my blog for a tutorial on doing this:
    http://soastuff.wordpress.com/2007/06/06/handling-opaque-datatypes-in-bpelesb/
    The only problem with this at the moment is that the BPEL function ora:parseEscapedXML has a bug in it - so it fails if the XML does not comply with the XSD it tries to convert to.
    This in turn causes an ORABPEL-09500 error which does not seem to get handled by the BPEL process so the process fails.
    As things stand, this bug remains - though Dave Berry has been very helpful in identifying it and bringing it to Muruga's attention. Hopefully we can get a fix (?) and a workaround whilst the fix is being implemented (?)
    Cheers.

  • Uk date format issue with ASP and Access Database

    I have an Asp form which updates records in an Access
    database. Problem is
    that the date format in the database record is dd/mm/yyyy
    (UK), when
    the record is displayed on the form it is mm/dd/yyyy(US)
    which after I
    update the record in the database the date has changed to the
    new format.
    I have tried everything I can to change the format but to no
    avail...anyone any ideas how I can resolve this issue?
    Thanks
    Steve

    stevo.s wrote:
    > Hi
    >
    > I have tried changing the format on the date field on
    the server behaviours
    > panelto ddmmyyy. Also have tried to set the form field
    format to ddmmyy. I have
    > also tried to use a function I got from a posting
    somehwere on the net to no
    > avail. <%function ddmmyyyy(varDate)
    > ddmmyyyy = Day(DateValue(varDate)) & "/" &
    Month(DateValue(varDate))
    > & "/" & Year(DateValue(varDate))
    > end function
    >
    > I believe that this is a recognised issue with
    Dreamweaver and Access but
    > can't seem to grasp the work around! Problem being I am
    teaching myself through
    > books and internet articles and can be weeks at a time
    without being able to
    > look at the issue..each time I come back to it it is
    like starting all over
    > again! I was hoping that somewhere out there there is a
    simple solution the
    > issue perhaps a date picker with the built in
    functionality to address the
    > issue...I am keen to understand how to deal with the
    issue rather than just
    > change my database date field to fudge the problem as I
    am in the UK and when I
    > eventually start to use the application I would like
    there to be some
    > consistency with dates and that users are familiar with
    the format.
    >
    > Any help gratefully received!
    Its not Dreamweaver, or Access, its your servers locale, its
    set to US
    format, not the UK.
    On your page at the top use:
    <% Session.LCID = 2057 %>
    This will force the page into using UK formatted dates. Use
    it on any
    page that needs to format the page correctly.
    Dooza
    Posting Guidelines
    http://www.adobe.com/support/forums/guidelines.html
    How To Ask Smart Questions
    http://www.catb.org/esr/faqs/smart-questions.html

  • Problem in getting data into database with standard direct input program

    HI All,
    I am having problem which is not updating the records in MM01 or MM02 with standard direct input program. i have data in internal table. from that table i am trying to upload into database by using background job MRP_MATERIAL_MASTER_DATA_LOAD.
    when i execute my program it is showing message job is started. then i go into sm37 and seethe job status by executing. there also i am seeing job completed succesfully.
    but if i go to mm03 and find the materials are updated or created there i couldn't find the material numbers which are from internal table.
    So if ny one help me it wil be great.
    Thanks in Advance
    Venkat N

    Hi Anil,
    Thanks for your answer, but i am facing problem is i have material no and denominator and Actual UOM and nominator field values in the flat file.....
    by using RMDATIND direct input program with MRP_MATERIAL_UPLOAD as job name for background job while uploading data into database.
    here i am not getting data in to database, but when i execute the job in sm37 it is showing that message job processing successfully completed...this is my status..
    if u can help me in this it will be gr8ful..
    Thanks,
    Venkat N

  • Data Integration issue with Adobe form

    Hi,
    We have been using many xACF forms in our WebDynpro java based application.
    For one of the xACF adobe form sometimes I'm not getting the data from adobe form to webdynpro context.
    Anyone can pls tell what may be the problem other then binding because that part i have already checked and it's working fine in some of the PC.
    Regards
    Ravindra Singh

    Hi Chintan,
    thanks for your reply.
    I tried that part already by converting xACF form to ZCI but facing issues with the drop-downs.
    Actually some of the dropdowns have values more than 1000. The value help for these drop downs comes as pop up in case of ZCI forms and the mazer problem is that in offline scenario doesn't even shows values for such dropdowns as a pop up or normal drop down.
    As per the SAP Note 1013227 we are using the enumerated type of drop downs for such larger values but no help in offline  scenario. I tried by changing the parameter value in visual admin as suggested in sap note which prevents pop-up for lager values but it's not recommended by SAP. Only because of this issue we are not able to migrate from xACF to ZCI.
    Do you have any idea how to get values for such drop downs in offline scenario.
    Thanks & Regards
    Ravindra Singh

  • Issue with generic extraction (Generic Delta)

    I have one more issue with Generic Delta Extraction. Here I selected the field related to my requirement but when I am trying to save I am getting error Still OLTP have errors.
    I need some help when we select generic delta wht are the settings. Like now I am working on Inventory Management.
    Please Some one help me as soon as possible.
    Thanks In advance.....
    Regds
    SDR.

    Hi ,
    I think there are some of the Keyfigure fields which needs 0Unit or 0Currency field to be present in the Extract Structure but they are not present . Check All unit of measure and Currency unit fields are not hidden .
    Regards,
    Vijay.

  • Xerces Validation Issues with Honour All schema Locations

    Way back in 2009/2010 there was a bug - 284272 - that was never updated with Xerces information about validation. We have a similar issue in Luna and Mars with the WTP validation framework (with Xerxes at the center) and I'm wondering if it's due to Xerces 2.9 or something else.
    I'm not a savvy XSD guy, but the issues detailed in 284272 still occur for us with certain schema combinations. We're encountering it specifically with the JBoss/SwitchYard project and the SCA schemas we extend for the SwitchYard runtime.
    The following error is displayed in the Problems tab:
    cvc-complex-type.2.4.a: Invalid content was found starting with element 'sca:interface.java'. One of '{"http://docs.oasis-open.org/ns/opencsa/sca/200912":documentation, "http://docs.oasis-open.org/ns/opencsa/sca/200912":interface, "http://docs.oasis-open.org/ns/opencsa/sca/200912":binding, "http://docs.oasis-open.org/ns/opencsa/sca/200912":callback, "http://docs.oasis-open.org/ns/opencsa/sca/200912":requires, "http://docs.oasis-open.org/ns/opencsa/sca/200912":policySetAttachment, "http://docs.oasis-open.org/ns/opencsa/sca/200912":extensions}' is expected. switchyard.xml src/main/resources/META-INF line 7 XML Problem
    Can someone revisit 284272 and let me know if this is indeed a Xerxes issue under the covers with the example included in the bug? It's not our example, but has the same general problem.
    I just don't want to open a duplicate of 284272 without knowing more about how the Honour All Schema Locations stuff works under the covers and if it is indeed caused by Xerces - indeed if it might be fixed by an update to Xerces further down the line (they're up to 2.11 and there was a 2.9.1 it looks like we might be able to more simply move to if it's fixed there).
    Thanks all!
    --Fitz

    I think this is the odata4 that you need for 7.3
    SCA
    ODATACXFEXT11_0-10012140.SCA
    SP11 for SAP ODATA4J+CXF-REST LIB 7.30
    0
    Info
    10352
    26.02.2014
    This might be the only package you need; it might be in your best interest to avoid manually downloading files and rely upon Solution Manager's MOpz feature to query your IdM system and let it track down all the needed packages.

  • Issue with Standard price

    The issue is coming when we are using currency as TND and HUF
    TND Currency has 3 decimal points
    HUF Currency has NO decimal points.
    1) Issue with TND Currency: - Price value shifting left side by one
    decimal
    Sample Scenario for TND:
    When a cart is getting created with quantity 100 EA and price per
    unit is 120.000 TND and when the user clicks on "Check",
    The price per unit is shifted to left side and the value is
    getting changed to 12.000 TND.
    And this impact is effecting on the Total Value also.
    12.000 TND * 100 EA = 1200.000 TND (But it should be 12000.000
    TND).
    This cart is approved with incorrect value and is available in
    the carryout sourcing.
    The price (Custom Price) of the shopping cart in the work list
    in sourcing is 12.000 TND (same as the changed value when the cart was
    created).
    When this cart is added to the buyer's work area, AGAIN standard
    SAP price field is getting changed from 12.000 TND to 1.200 TND.
    Changes happen like this 120.000 TND -> 12.000 TND -> 1.200 TND.
    2) Issue with HUF Currency: - 2 zeros adding into right side.
    Sample Scenario for HUF:
    When a cart is getting created with quantity 100 EA and price per
    unit is 120 HUF and when the user clicks on "Check",
    The price per unit getting updated as 12000 HUF.(Please note that
    2 zeros are added in right side and there is no decimal point for this
    currency)
    And this impact is effecting on the Total Value also .
    12000* 100 EA = 1200000 (But it should be 120 * 100 EA = 12000)
    This cart is approved and is available in carryout sourcing.
    The price (Custom Price) of the shopping cart in the work list
    in sourcing is 12000 (same as the changed value when the cart was
    created).
    When this cart is added to the buyer's work area, again
    standard SAP price field is getting changed from 12000 to 1200000 (Its
    Already changed once
    from 120 to 12000 while creating the shopping cart ) in
    buyer's sourcing work area
    Changes happen like this 120 HUF -> 12000 HUF -> 1200000 HUF.
    In case anybody has worked on a similar scenario before, please provide some pointers in this area.
    Regards,
    Bharathi

    Hi. Have a look for applicable OSS notes for function module CONVERT_TO_LOCAL_CURRENCY.
    We have had numerous issues with currencies, and it is normally this function that causes the problems.
    Regards,
    Dave.

  • Unicode and non-unicode string data types Issue with 2008 SSIS Package

    Hi All,
    I am converting a 2005 SSIS Package to 2008. I have a task which has SQL Server as the source and Oracle as the destination. I copy the data from a SQL server view with a field nvarchar(10) to a field of a oracle table varchar(10). The package executes fine
    on my local when i use the data transformation task to convert to DT_STR. But when I deploy the dtsx file on the server and try to run from an SQL Job Agent it gives me the unicode and non-unicode string data types error for the field. I have checked the registry
    settings and its the same in my local and the server. Tried both the data conversion task and Derived Column task but with no luck. Pls suggest me what changes are required in my package to run it from the SQL Agent Job.
    Thanks.

    What is Unicode and non Unicode data formats
    Unicode : 
    A Unicode character takes more bytes to store the data in the database. As we all know, many global industries wants to increase their business worldwide and grow at the same time, they would want to widen their business by providing
    services to the customers worldwide by supporting different languages like Chinese, Japanese, Korean and Arabic. Many websites these days are supporting international languages to do their business and to attract more and more customers and that makes life
    easier for both the parties.
    To store the customer data into the database the database must support a mechanism to store the international characters, storing these characters is not easy, and many database vendors have to revised their strategies and come
    up with new mechanisms to support or to store these international characters in the database. Some of the big vendors like Oracle, Microsoft, IBM and other database vendors started providing the international character support so that the data can be stored
    and retrieved accordingly to avoid any hiccups while doing business with the international customers.
    The difference in storing character data between Unicode and non-Unicode depends on whether non-Unicode data is stored by using double-byte character sets. All non-East Asian languages and the Thai language store non-Unicode characters
    in single bytes. Therefore, storing these languages as Unicode uses two times the space that is used specifying a non-Unicode code page. On the other hand, the non-Unicode code pages of many other Asian languages specify character storage in double-byte character
    sets (DBCS). Therefore, for these languages, there is almost no difference in storage between non-Unicode and Unicode.
    Encoding Formats: 
    Some of the common encoding formats for Unicode are UCS-2, UTF-8, UTF-16, UTF-32 have been made available by database vendors to their customers. For SQL Server 7.0 and higher versions Microsoft uses the encoding format UCS-2 to store the UTF-8 data. Under
    this mechanism, all Unicode characters are stored by using 2 bytes.
    Unicode data can be encoded in many different ways. UCS-2 and UTF-8 are two common ways to store bit patterns that represent Unicode characters. Microsoft Windows NT, SQL Server, Java, COM, and the SQL Server ODBC driver and OLEDB
    provider all internally represent Unicode data as UCS-2.
    The options for using SQL Server 7.0 or SQL Server 2000 as a backend server for an application that sends and receives Unicode data that is encoded as UTF-8 include:
    For example, if your business is using a website supporting ASP pages, then this is what happens:
    If your application uses Active Server Pages (ASP) and you are using Internet Information Server (IIS) 5.0 and Microsoft Windows 2000, you can add "<% Session.Codepage=65001 %>" to your server-side ASP script.
    This instructs IIS to convert all dynamically generated strings (example: Response.Write) from UCS-2 to UTF-8 automatically before sending them to the client.
    If you do not want to enable sessions, you can alternatively use the server-side directive "<%@ CodePage=65001 %>".
    Any UTF-8 data sent from the client to the server via GET or POST is also converted to UCS-2 automatically. The Session.Codepage property is the recommended method to handle UTF-8 data within a web application. This Codepage
    setting is not available on IIS 4.0 and Windows NT 4.0.
    Sorting and other operations :
    The effect of Unicode data on performance is complicated by a variety of factors that include the following:
    1. The difference between Unicode sorting rules and non-Unicode sorting rules 
    2. The difference between sorting double-byte and single-byte characters 
    3. Code page conversion between client and server
    Performing operations like >, <, ORDER BY are resource intensive and will be difficult to get correct results if the codepage conversion between client and server is not available.
    Sorting lots of Unicode data can be slower than non-Unicode data, because the data is stored in double bytes. On the other hand, sorting Asian characters in Unicode is faster than sorting Asian DBCS data in a specific code page,
    because DBCS data is actually a mixture of single-byte and double-byte widths, while Unicode characters are fixed-width.
    Non-Unicode :
    Non Unicode is exactly opposite to Unicode. Using non Unicode it is easy to store languages like ‘English’ but not other Asian languages that need more bits to store correctly otherwise truncation will occur.
    Now, let’s see some of the advantages of not storing the data in Unicode format:
    1. It takes less space to store the data in the database hence we will save lot of hard disk space. 
    2. Moving of database files from one server to other takes less time. 
    3. Backup and restore of the database makes huge impact and it is good for DBA’s that it takes less time
    Non-Unicode vs. Unicode Data Types: Comparison Chart
    The primary difference between unicode and non-Unicode data types is the ability of Unicode to easily handle the storage of foreign language characters which also requires more storage space.
    Non-Unicode
    Unicode
    (char, varchar, text)
    (nchar, nvarchar, ntext)
    Stores data in fixed or variable length
    Same as non-Unicode
    char: data is padded with blanks to fill the field size. For example, if a char(10) field contains 5 characters the system will pad it with 5 blanks
    nchar: same as char
    varchar: stores actual value and does not pad with blanks
    nvarchar: same as varchar
    requires 1 byte of storage
    requires 2 bytes of storage
    char and varchar: can store up to 8000 characters
    nchar and nvarchar: can store up to 4000 characters
    Best suited for US English: "One problem with data types that use 1 byte to encode each character is that the data type can only represent 256 different characters. This forces multiple
    encoding specifications (or code pages) for different alphabets such as European alphabets, which are relatively small. It is also impossible to handle systems such as the Japanese Kanji or Korean Hangul alphabets that have thousands of characters."<sup>1</sup>
    Best suited for systems that need to support at least one foreign language: "The Unicode specification defines a single encoding scheme for most characters widely used in businesses around the world.
    All computers consistently translate the bit patterns in Unicode data into characters using the single Unicode specification. This ensures that the same bit pattern is always converted to the same character on all computers. Data can be freely transferred
    from one database or computer to another without concern that the receiving system will translate the bit patterns into characters incorrectly.
    https://irfansworld.wordpress.com/2011/01/25/what-is-unicode-and-non-unicode-data-formats/
    Thanks Shiven:) If Answer is Helpful, Please Vote

  • Data load issue with export data source - BW 3.5

    Hi,
    We are facing issues in loading data with the help of export data source.
    We have created export data source of 0PCA_C01 cube. With the help of this export datasource,  we are loading data to other custom cube. Scenario is working fine in development server.
    But when we transported objects to quality server data is not getting loaded to custom target cube.
    It is extracting zero records.  All transports are ok and we have generated export datasource in quality before transports .Also regenerated export datasource after transport and activated infosource, update rule via RS* programs.  Every object is active but data is not getting extracted.
    RSA3 for 80PCA_C01 datasource isn't extracting any record in Quality. Records getting extracted in development.   We are in BW 3.5 with patch level 19.
    Please guide us to resolve the issue.
    Thanks,
    Aditya

    Hi
    Make sure that you have relevant Role & Authorization at Quality/PRS.
    You have to Transport the Source Cube first and then Create a Generate Export Data Source in QAS. Then, replicate data sources for BW QAS Soruce System. Make sure this replicated Data Source in QAS. Only then can transport new update rules for second cube.
    Hope it helps and clear

  • Generic Data source Issue with Infoset

    Hi Experts,
    I have created generic data source with Infoset.As per customer requirement, I have to modify join condition between tables from equal to left outer join in infoset.I have tested scenario successfully using infoset query. I have done below steps in generic data source to check the same result.
    1.Modified Join condition in infoset and generated the infoset again.
    2. Generated generic datasource in RSO2.
    When I check same scenario in RSA3, I am getting 0 records.Please help me out.
    Thanks in Advance.
    Regards,
    Raju

    Thanks Sriram for your reply.
    I am using same user id for both SQ02 and RSA3.
    I have Two infosets with same join conditions.I have created SQVI report with First infoset and Generic data source with second infoset.
    I have changed join condition second infoset which is used for generic data source.I am still getting same number of records in both SQVI report and in RSA3.I have validated records by considering few records to cross check again.
    The change in join condition doesnt show any impact in changing the number of recrods.
    I  have activated extraction structure in RSA6 but still getting same records.
    Thanks,
    Raju

  • JBO-JHS-00114: JHS-00114: Unknown date validation issue

    Hi,
    I am facing a strange issue while filling date field values in certain pages of application.
    I have assigned Date Format to     'MM/dd/yyyy' at Service level in Jheadstart app structure definition .I have custom datefield velocity template to display hint along with data field.I assigned this to FORM_DATE_FIELD template.
    Now I generated pages in my application. All pages works as expected except couple of pages.Even though I enter valid value in date field, It throws me an error "JBO-JHS-00114: JHS-00114: Date 09/04/2007 must be of format dd-MMM-yyyy ".
    I could not quite understand where it is picking up this special format?
    I went through generated JSPX file code, it is fetching from nls.datepattern.This code is same across all pages.Then where is this error coming from?
    Thanks.

    Here is custom date velocity template,
    <af:selectInputDate #ITEM_ID_IN_FORM() #ITEM_VALUE_IN_FORM() #ITEM_PROMPT_IN_FORM()
    #ITEM_PARTIAL_TRIGGERS() #ITEM_REQUIRED_IN_FORM()
    #ITEM_READ_ONLY_IN_FORM() #ITEM_RENDERED_IN_FORM()
    #ITEM_DISABLED_IN_FORM() tip="(mm/dd/yyyy)" #DEPENDS_ON_ITEM_PROPS_FORM()>
    <af:convertDateTime pattern="#{nls.datepattern}"/>
    </af:selectInputDate>

Maybe you are looking for