Nchar issue

Hi we're having an issue on our production server with nchars. It is as follows:
We have a person table with an active field which has an nchar with Y/N check constraint on it. We also have a view which includes all the fields of the person table and some information from a few other tables.
The issue we have is a bunch of queries on the person_view which look for active people return no rows on our production server but work fine on our test server. We have tracked this down to the active field. (Please see example queries below the result in bold is the cause of our issue). I believe the problem is that the test server has a different character set to the production server (not a good idea I know). I am currently trying to rectify this however it has not be straight forward.
What's really weird is that the problem seems limited to views and does not occur when querying a table. Can anyone explain why this occurs? Is this a bug or is there a good reason for this behavior?
Test Server:_
SQL> select count(*) from person where active = 'Y';
COUNT(*)
519
SQL> select count(*) from person_view where active = 'Y';
COUNT(*)
519
SQL> select count(*) from person_view where active like 'Y';
COUNT(*)
519
SQL> SELECT value$ FROM sys.props$ WHERE name = 'NLS_CHARACTERSET' ;
VALUE$
AL32UTF8
Production Server:*
SQL> select count(*) from person where active = 'Y';
COUNT(*)
519
SQL> select count(*) from person_view where active = 'Y';
COUNT()*
*0*
SQL> select count(*) from person_view where active like 'Y';
COUNT(*)
519
SQL> select value$ from sys.props$ where name = 'NLS_CHARACTERSET';
VALUE$
WE8MSWIN1252

Details below. (Test and production servers quite different which I know is bad practice. Am working to rectify this also.)
Thanks in advance.
Test Server:
O/S = MS Server 2008 Standard Service Pack 1 32 bit
Oracle 10g XE (10.2.0.1.0)
SQL> select * from nls_database_parameters;
PARAMETER
VALUE
NLS_LANGUAGE
AMERICAN
NLS_TERRITORY
AMERICA
NLS_CURRENCY
$
PARAMETER
VALUE
NLS_ISO_CURRENCY
AMERICA
NLS_NUMERIC_CHARACTERS
NLS_CHARACTERSET
AL32UTF8
PARAMETER
VALUE
NLS_CALENDAR
GREGORIAN
NLS_DATE_FORMAT
DD-MON-RR
NLS_DATE_LANGUAGE
AMERICAN
PARAMETER
VALUE
NLS_SORT
BINARY
NLS_TIME_FORMAT
HH.MI.SSXFF AM
NLS_TIMESTAMP_FORMAT
DD-MON-RR HH.MI.SSXFF AM
PARAMETER
VALUE
NLS_TIME_TZ_FORMAT
HH.MI.SSXFF AM TZR
NLS_TIMESTAMP_TZ_FORMAT
DD-MON-RR HH.MI.SSXFF AM TZR
NLS_DUAL_CURRENCY
$
PARAMETER
VALUE
NLS_COMP
BINARY
NLS_LENGTH_SEMANTICS
BYTE
NLS_NCHAR_CONV_EXCP
FALSE
PARAMETER
VALUE
NLS_NCHAR_CHARACTERSET
AL16UTF16
NLS_RDBMS_VERSION
10.2.0.1.0
20 rows selected.
Production Server:
O/S = MS Server 2008 Standard Service Pack 1 64 bit
Oracle 11g Standard Edition (11.1.0.7.0)
SQL> select * from nls_database_parameters;
PARAMETER VALUE
NLS_LANGUAGE AMERICAN
NLS_TERRITORY AMERICA
NLS_CURRENCY $
NLS_ISO_CURRENCY AMERICA
NLS_NUMERIC_CHARACTERS .,
NLS_CHARACTERSET WE8MSWIN1252
NLS_CALENDAR GREGORIAN
NLS_DATE_FORMAT DD-MON-RR
NLS_DATE_LANGUAGE AMERICAN
NLS_SORT BINARY
NLS_TIME_FORMAT HH.MI.SSXFF AM
PARAMETER VALUE
NLS_TIMESTAMP_FORMAT DD-MON-RR HH.MI.SSXFF AM
NLS_TIME_TZ_FORMAT HH.MI.SSXFF AM TZR
NLS_TIMESTAMP_TZ_FORMAT DD-MON-RR HH.MI.SSXFF AM TZR
NLS_DUAL_CURRENCY $
NLS_COMP BINARY
NLS_LENGTH_SEMANTICS BYTE
NLS_NCHAR_CONV_EXCP FALSE
NLS_NCHAR_CHARACTERSET AL16UTF16
NLS_RDBMS_VERSION 11.1.0.7.0
20 rows selected.

Similar Messages

  • NCHAR issue with oracle database using JDBC adapter

    Hi,
    We have a requirement to develop an XI interface from FTP server(File adapter) to oracle database using JDBC adapter. In the oracle database table few fields are of type NCHAR/NVARCHAR. when we try to insert the character(A,B,c..) values into oracle table fields of type NCHAR/NVARCHAR, we are getting the following error message in the JDBC adapter audit log. IF we pass the numeric value to the same field, then we are able to insert the records successfully.
    Unable to execute statement for table or stored procedure. 'IPCSDD_DOWNLOAD_PROCESS' (Structure 'StatementName1') due to java.sql.SQLException: ORA-00904: "P": invalid identifier
    2010-10-19 22:29:59 Error JDBC message processing failed; reason Error processing request in sax parser: Error when executing statement for table/stored proc. 'IPCSDD_DOWNLOAD_PROCESS' (structure 'StatementName1'): java.sql.SQLException: ORA-00904: "P": invalid identifier
    2010-10-19 22:29:59 Error MP: Exception caught with cause com.sap.aii.af.ra.ms.api.RecoverableException: Error processing request in sax parser: Error when executing statement for table/stored proc. 'IPCSDD_DOWNLOAD_PROCESS' (structure 'StatementName1'): java.sql.SQLException: ORA-00904: "P": invalid identifier
    Please find the system information below.
    Oracle version- 10.2.4
    XI version - 3.0/ service pack 19
    JDBC driver- oracle.jdbc.driver.OracleDriver
    Please suggest.
    Thanks,
    Venkata
    Edited by: Venkata Narayana  Eepuri on Oct 21, 2010 12:10 AM

    Dear Venkata Narayana,
    Concerning the error, kindly go through the following note :
    731 - Collective note: ORA-00904
    follow the recommendations mentioned in that and please check if that helps.
    Best Regards
    Nishwanth

  • Oracle 10g db character set issue

    I have a database 10g with database character set western
    european "WE8ISO8859P1" and we are receiving data from source
    database with database character set as "UTF8" during data load
    for one of the tables we receive the following error "ORA-29275:
    partial multibyte character" I understand this might be due to
    the fact western european character set is not a subset/superset
    of UTF8 .Am i right ? What would be the way around this ?

    It is certainly possible that the issue is that your database characterset is a subset of UTF8.
    How are you getting the data? Are we talking about a flat file? A query over a database link? Something else?
    Does the data you're getting contain characters that cannot be represented in the ISO-8859 1 character set? It is quite common to send UTF-8 encoded files even when the underlying data is representable in other 8-bit character sets (like ISO-8859 1).
    What are you trying to do with the data? Are you trying to load it into a CHAR/ VARCHAR2 column? A CLOB? A BLOB? An NCHAR/ NVARCHAR2? Something else?
    Justin

  • Error ORA-00980: synonym translation is no longer valid when i issue

    i accidentally drop one of my table. To bring it back i used Imp tool. But it gives an error ORA-00980: synonym translation is no longer valid when i issue
    Select * from ship_sched_fact_notice_det
    Action taken using Imp
    Username: erpdada
    Password:
    Connected to: Oracle Database 10g Release 10.1.0.2.0 - Production
    Import file: EXPDAT.DMP > e:\erpdada_august13.dmp
    Enter insert buffer size (minimum is 8192) 30720> 30720
    Export file created by EXPORT:V10.01.00 via conventional path
    Warning: the objects were exported by SYSTEM, not by you
    import done in WE8MSWIN1252 character set and AL16UTF16 NCHAR character set
    import server uses AL32UTF8 character set (possible charset conversion)
    List contents of import file only (yes/no): no > y
    Import entire export file (yes/no): no > n
    Username: erpdada
    Enter table(T) or partition(T:P) names. Null list means all tables for user
    Enter table(T) or partition(T:P) name or . if done: ship_sched_fact_notice_det
    Enter table(T) or partition(T:P) name or . if done: .
    "CREATE PUBLIC SYNONYM "XMLTYPE" FOR "SYS"."XMLTYPE""
    "CREATE TABLE "SHIP_SCHED_FACT_NOTICE_DET" ("SSFND_ITEM_NOT_DET_ID" NUMBER, "
    ""SSFN_FACTORY_ID" VARCHAR2(40) NOT NULL ENABLE, "SSFN_FAC_NOTICE_ID" NUMBER"
    " NOT NULL ENABLE, "SSD_SHIP_ITEM_ID" NUMBER NOT NULL ENABLE, "OI_ITEM_ID" N"
    "UMBER NOT NULL ENABLE, "ORDER_TYPE" VARCHAR2(1), "SSFND_ITEM_CODE" VARCHAR2"
    "(300), "SSFND_QTY" NUMBER, "SSFND_CTN" NUMBER, "SSFND_TOT_CBM" NUMBER, "SSF"
    "ND_REM1" VARCHAR2(256), "SSFND_REM2" VARCHAR2(256), "SSFND_UNIT_ID" VARCHAR"
    "2(40), "SSFND_IS_EXCL_SUMMARY" VARCHAR2(1))  PCTFREE 10 PCTUSED 40 INITRANS"
    " 1 MAXTRANS 255 STORAGE(INITIAL 81920 FREELISTS 1 FREELIST GROUPS 1 BUFFER_"
    "POOL DEFAULT) TABLESPACE "ORDER_DATA01" LOGGING NOCOMPRESS"
    . . skipping table "SHIP_SCHED_FACT_NOTICE_DET"
    "GRANT DELETE ON "SHIP_SCHED_FACT_NOTICE_DET" TO "SHIPMENT_NOTICE""
    "GRANT INSERT ON "SHIP_SCHED_FACT_NOTICE_DET" TO "SHIPMENT_NOTICE""
    "GRANT SELECT ON "SHIP_SCHED_FACT_NOTICE_DET" TO "SHIPMENT_NOTICE""
    "GRANT UPDATE ON "SHIP_SCHED_FACT_NOTICE_DET" TO "SHIPMENT_NOTICE""
    "CREATE UNIQUE INDEX "PK_SHIP_SCHED_FACT_NOTICE_DET" ON "SHIP_SCHED_FACT_NOT"
    "ICE_DET" ("SSFND_ITEM_NOT_DET_ID" )  PCTFREE 10 INITRANS 2 MAXTRANS 255 STO"
    "RAGE(INITIAL 65536 FREELISTS 1 FREELIST GROUPS 1 BUFFER_POOL DEFAULT) TABLE"
    "SPACE "ORDER_DATA01" LOGGING"
    "ALTER TABLE "SHIP_SCHED_FACT_NOTICE_DET" ADD  CONSTRAINT "PK_SHIP_SCHED_FAC"
    "T_NOTICE_DET" PRIMARY KEY ("SSFND_ITEM_NOT_DET_ID") USING INDEX PCTFREE 10 "
    "INITRANS 2 MAXTRANS 255 STORAGE(INITIAL 65536 FREELISTS 1 FREELIST GROUPS 1"
    " BUFFER_POOL DEFAULT) TABLESPACE "ORDER_DATA01" LOGGING ENABLE "
    "ALTER TABLE "SHIP_SCHED_FACT_NOTICE_DET" ADD CONSTRAINT "SHIP_SCHED_FACT_NO"
    "TICE_DE_FK1" FOREIGN KEY ("SSD_SHIP_ITEM_ID") REFERENCES "SHIPPING_SCHEDULE"
    "_DETAILS" ("SSD_SHIP_ITEM_ID") ENABLE NOVALIDATE"
    "ALTER TABLE "SHIP_SCHED_FACT_NOTICE_DET" ADD CONSTRAINT "SHIP_SCHED_FACT_NO"
    "TICE_DE_FK2" FOREIGN KEY ("SSFN_FAC_NOTICE_ID") REFERENCES "SHIP_SCHED_FACT"
    "ORY_NOTICE" ("SSFN_FAC_NOTICE_ID") ENABLE NOVALIDATE"
    "ALTER TABLE "SHIP_SCHED_FACT_NOTICE_DET" ENABLE CONSTRAINT "SHIP_SCHED_FACT"
    "_NOTICE_DE_FK1""
    "ALTER TABLE "SHIP_SCHED_FACT_NOTICE_DET" ENABLE CONSTRAINT "SHIP_SCHED_FACT"
    "_NOTICE_DE_FK2""
    Import terminated successfully with warnings.
    I drop the old synonym and re-create it but still it gives the same error...

    Since you are on 10g did you try flashback table feature - http://tonguc.wordpress.com/2007/01/01/oracle-10g-flashback-versions-query-drop-table-and-recyclebin-management/
    FLASHBACK TABLE <tabname> TO BEFORE DROP;
    If you didnt use purge option, the dropped object is on your recyclebin - http://download-uk.oracle.com/docs/cd/B19306_01/server.102/b14200/statements_9018.htm#sthref9593
    Best regards.

  • Excel import error with NCHAR field on v1.2.1

    Hi, we've tried to import data from Excel to an Oracle table (10g) using SQL Developer 1.2.1. However it failed with the following error in Log Window:
    ORA-00917: missing comma
    Further checking revealed that one field in Oracle table was defined as NCHAR, while the generated DML value was without quote:-
    insert into test_table (field1,field2) VALUES(123,1B);
    When we replaced the '1B' value to numeric value 11, the import worked. As we're using an application package, we cannot simply modify the table/field's attribute from NCHAR to CHAR. Is there a way to configure SQL Developer that will have the quote placed like:-
    insert into test_table (field1,field2) VALUES(123,'1B');
    which works fine in SQL Developer Query mode.
    Please advise, thanks :)

    Note and thanks for handling and fixing this issue.
    Furthermore to the import function, we also noted that the default setting cannot fit to inserting data with Chinese (or Double Byte) characters, we will need to add a "N" as a prefix, e.g.:-
    insert into table_a (field_a, field_b) values(123, N'一二三')
    If "N" is missing, field_b will only contains garbage.
    As we notice that the import will generate insert statements, please see if you can include this problem as part of the "missing quote" fix, thanks.

  • Oracle 11g Export issue

    I have Oracle 11g (11.1.0.6) in one server, and I have 10g (10.2.0.3) database in all other server. I am trying to do EXP (normal export not expdp) of my 10g database using the 11g client. I am getting the following error.
    Export: Release 11.1.0.6.0 - Production on Wed Jul 22 15:17:03 2009
    Copyright (c) 1982, 2007, Oracle.  All rights reserved.
    Connected to: Oracle Database 10g Enterprise Edition Release 10.2.0.3.0 - Production
    With the Partitioning, OLAP and Data Mining options
    Export done in US7ASCII character set and AL16UTF16 NCHAR character set
    server uses WE8MSWIN1252 character set (possible charset conversion)
    EXP-00008: ORACLE error 1003 encountered
    ORA-01003: no statement parsed
    . . exporting table SNP_DT
    EXP-00008: ORACLE error 904 encountered
    ORA-00904: "MAXSIZE": invalid identifier
    EXP-00008: ORACLE error 942 encountered
    ORA-00942: table or view does not exist
    EXP-00024: Export views not installed, please notify your DBA
    EXP-00000: Export terminated unsuccessfully
    Is there any way way to solve this. Could some one please help me.
    Thanks
    Mano
    Edited by: Mano Rangasamy on Jul 22, 2009 3:24 PM
    Edited by: Mano Rangasamy on Jul 22, 2009 3:26 PM

    I have the same issue, but my client (export) is 11.1.0.7.0 and my database is 11.1.0.6.0. I beleaved that the same versions (11) have to work toghether. But they do not.
    Does it mean that I have to reinstall my client to have en export that works?

  • BUG: retrieving NCHAR value using ViewObject returns "???" with OCI driver

    Hi,
    after upgrading JDeveloper to 10.1.3.1 we cannot use OCI JDBC drivers, because they don't work correctly - characters, that do not exist in database character set (EE8ISO8859P2 in our case) are converted to "?". Thin driver works well. We need to use OCI, because we had some issues with RAC with thin driver.
    Test case is easy: create table with NCHAR column, populate with unicode data, create ViewObject to select from this table and it will work only with THIN and not with OCI driver. This worked well with 10.1.3.0. When I tried to directly connect to DB and select that table using java.sql.Statement, it worked with both drivers, that's why I suspect the mechanics inside ADF BC.
    I logged a SR for this in Metalink, but try this way as well, for someone else could hit the same problem.
    Thanks for any help,
    Viliam

    Hi,
    after upgrading JDeveloper to 10.1.3.1 we cannot use OCI JDBC drivers, because they don't work correctly - characters, that do not exist in database character set (EE8ISO8859P2 in our case) are converted to "?". Thin driver works well. We need to use OCI, because we had some issues with RAC with thin driver.
    Test case is easy: create table with NCHAR column, populate with unicode data, create ViewObject to select from this table and it will work only with THIN and not with OCI driver. This worked well with 10.1.3.0. When I tried to directly connect to DB and select that table using java.sql.Statement, it worked with both drivers, that's why I suspect the mechanics inside ADF BC.
    I logged a SR for this in Metalink, but try this way as well, for someone else could hit the same problem.
    Thanks for any help,
    Viliam

  • Data Conversion and Derived Column issues

    I have a strange issue occurring with one of my SSIS packages. 
    What needs to happen:
    I have to read data from a table that stores a field in NCHAR(40)
    Send it through a Data Cleansing SSIS Component that forcefully outputs the data at a cleansed state as DT_STR(50)
    Update the same source table with cleansed data - Using an UPSERT third party tool
    Of course, I can't update the tables that stores data in NCHAR(40) with data from DT_STR(50), so I'm trying to use the Data Conversion Component, the Derived Column Component or a combination of, to set the data to DT_WSTR as well as to set the correct length,
    from 50 to 40.
    The Data Conversion Component fails when I try to set the incoming data (DT_STR(50)) to DT_WSTR(40):
    [Data Conversion [186]] Error: Data conversion failed while converting column "MD_Address1" (97) to column "_MD_Address1" (190).  The conversion returned status value 2 and status text
    "The value could not be converted because of a potential loss of data.".
    Now I tried the same thing with the Derived Column Component, converting the data from the DQS component from (DT_STR(50)) to DT_WSTR(40) and there's the error message:
    [Derived Column [196]] Error: The "Derived Column" failed because truncation occurred, and the truncation row disposition on "Derived Column.Inputs[Derived Column Input].Columns[_MD_Address1]"
    specifies failure on truncation. A truncation error occurred on the specified object of the specified component.
    I also tried a combination of Data Conversion (From DT_STR(50) to DT_WSTR(50)) and a Derived Column just casting to the correct size, but that didn't work either. 
    Every time I try to change the size, one of the components fail. 
    It appears that my issue has nothing to do with the data types, but the actual data length.
    Why can't SSIS CAST my input from 50 to 40?
    What do I need to do to cast/convert data from DT_STR(50) to DT_WSTR(40)?

    Hi IgorSantos_FL,
    It is the expected behavior when we try to convert DT_STR(50) (means maximum 50 characters in the value) to DT_WSTR(40) (means maximum 40 characters in the value). However, the truncation issue should not occur if you convert DT_STR(50) to
    DT_WSTR(50). Could you post the error message that you received when converting DT_STR(50) to DT_WSTR(50)? It may be a different issue.
    Regards,
    Mike Yin
    TechNet Community Support

  • Transportable Tablespace Issue

    Hi all,
    I dont know somebody else has posted this error here before, I am sorry If so..
    I am getting an error when I am transporting tablespace's metadata... can anyone help me about this issue ?
    My OS is RH 4
    ======================================================================
    [oracle@localhost export]$ exp parfile=/home/oracle/oracle9i/backup/export/prams.dat
    Export: Release 9.2.0.1.0 - Production on Mon Apr 28 13:33:23 2008
    Copyright (c) 1982, 2002, Oracle Corporation. All rights reserved.
    Username: sys as sysdba
    Password:
    Connected to: Oracle9i Enterprise Edition Release 9.2.0.1.0 - Production
    With the Partitioning, OLAP and Oracle Data Mining options
    JServer Release 9.2.0.1.0 - Production
    Export done in US7ASCII character set and AL16UTF16 NCHAR character set
    Note: table data (rows) will not be exported
    About to export transportable tablespace metadata...
    EXP-00008: ORACLE error 1001 encountered
    ORA-01001: invalid cursor
    ORA-06512: at "SYS.DBMS_SYS_SQL", line 819
    ORA-06512: at "SYS.DBMS_SQL", line 19
    ORA-06512: at "SYS.DBMS_TTS", line 774
    ORA-01187: cannot read from file 201 because it failed verification tests
    ORA-01110: data file 201: '/home/oracle/oracle9i/oradata/db1/dbf/temp01.dbf'
    ORA-06512: at "SYS.DBMS_PLUGTS", line 1093
    ORA-06512: at line 1
    EXP-00000: Export terminated unsuccessfully
    ====================================================================
    My parameter file looks like below
    TRANSPORT_TABLESPACE=y
    TABLESPACES=users
    FILE=/home/oracle/oracle9i/backup/export/ts_users_28042008_10:04.dmp
    LOG=/home/oracle/oracle9i/backup/export/explog.log

    Check the similar sort of thread.
    Re: Transportable tablespace
    Regards,
    Sabdar Syed.

  • SQl Server - cant compare with values in table field have nchar as datatype

    Hello all!
    I am using jtds driver to connect with SQL server. My problem is i am saving arabic text into the SQL Server. The fields which contains arabic text have the datatype nchar. I can insert the values by simply placing N before the actual arabic text which i am reading from a file.
    I mean insert into [table] values(N'some arabic text') works fine.
    But when i try to select the values from the table and want to pass some paramerts (in arabic text) in where clause it doesnt return me anything.
    e.g select * from [table] where [column(nchar type)] = N'arabic text(read from file)' return nothing. Result set is always empty.
    Queryis fine as i printed the query in seperate file and run it on Query analyzer which gives me the right results.
    Any suggestions?? Please i am dying conditions tell me any solutions immediately
    Regards,
    Alam sher

    String url = "............";
    Properties props = new Properties();
    props.put("charset","cp1256");
    Connection con = DriverManager.getConnection(url,props);now this way if i get the resultset it would return me the expected resultset:)
    To compare an arabic word against the values in result set one also has to convert it into cp1256 encoding line this.
    new String([arabic word].getBytes(), "cp1256");
    The prolem was not to print the arabic words on console basically. No doubt on console it would print ???? as it doesnot support this encoding. Originial issu was comparing a string containig the arabic text with the values returned by the result set.
    This was the solution i could figure out and its working.
    Thanks anyways.
    Regards,
    Alam Sher

  • Import issue using imp!

    Hi all,
    I am trying to import using imp and I am getting this error while import.
    import done in US7ASCII character set and AL16UTF16 NCHAR character set
    import server uses WE8MSWIN1252 character set (possible charset conversion)
    export client uses UTF8 character set (possible charset conversion)what is the imp script which resolves this issue.
    Thanks in advance
    sampoorna

    the imp script which resolves this issue.Its not an error, its a warning. Set the NLS_LANG environment variable appropriately. Appears the export environment was from a UTF8 database and the import client settings may cause improper character translations.
    If the target database character set is also UTF8 you might get away with setting NLS_LANG for a UTF8. And see support note 179133.1 "The correct NLS_LANG in a Windows Environment".

  • New DVR Issues (First Run, Channel Switching, etc.)

    I've spent the last 30 minutes trying to find answers through the search with no luck, so sorry if I missed something.
    I recently switched to FIOS from RCN cable in New York.  I've gone through trying to setup my DVR and am running into issues and was hoping for some answers.
    1.  I setup two programs to record at 8PM, I was watching another channel at the time and only half paying attention.  Around 8:02 I noticed a message had popped up asking if I would like to switch channels to start recording.  I was expecting it to force it to switch like my old DVR, but in this case it didn't switch and I missed the first two minutes of one of the shows.  I typically leave my DVR on all day and just turn off the TV, this dual show handling will cause issues with that if I forget to turn off the DVR.  Is there a setting I can change that will force the DVR to choose one of the recording channels?
    2.  I setup all my recordings for "First Run" because I only want to see the new episodes.  One show I setup was The Daily Show on comedy central, which is shown weeknights at 11pm and repeated 3-4 times throughout the day.  My scheduled recordings is showing all these as planned recordings even though only the 11pm show is really "new".  Most of the shows I've setup are once a week so they aren't a problem, but this seems like it will quickly fill my DVR.  Any fixes?
    Thanks for the help.
    Solved!
    Go to Solution.

    I came from RCN about a year ago.  Fios is different in several ways, not all of them desirable.  Here are several ways to get--and fix--unwanted recordings from a series recording setup.
    Some general principles. 
    Saving changes.  When you originally create a series with options, or if you go back to edit the options for an existing series, You MUST save the Series Options changes.  Pretty much everywhere else in the user interface, when you change an option, the change takes effect immediately--but not in Series Options.  Look at the Series Options window.  Look at the far right side.  There is a vertical "Save" bar, which you must navigate to and click OK on to actually save your changes.  Exiting the Series Options window without having first saved your changes loses all your attempted changes--immediately.
    Default Series Options.  This is accessed  from [Menu]--DVR--Settings--Default Series Options.  This will bring up the series options that will automatically be applied to the creation of a NEW series. The options for every previously created series will not be affected by a subsequent modification of the Default Series Options.  You should set these options to the way you would like them to be for the majority of series recordings that you are likely to create.  Be sure to SAVE your changes.  This is what you will get when you select "Create Series Recording" from the Guide.  When creating a new series recording where you think that you may want options different from the default, select "Create Series with Options" instead.  Series Options can always be changed for any individual series set up later--but not for all series at once.
    Non-series recordings.  With Fios you have no directly available options for these.  With RCN and most other DVRs, you can change the start and end times for individual episodes, including individual episodes that are also in a series.  With Fios, your workarounds are to create a series with options for a single program, then delete the series later;  change the series options if the program is already in a series, then undo the changes you made to the series options later; or schedule recordings of the preceding and/or following shows as needed.
    And now, to the unwanted repeats. 
    First, make sure your series options for the specific series in question--and not just the series default options--include "First Run Only".  If not, fix that and SAVE.  Then check you results by viewing the current options using the Series Manager app under the DVR menu.
    Second, and most annoying, the Guide can have repeat programs on your channel tagged as "New".  It happens.  Set the series option "Air Time" to "Selected Time".  To make this work correctly, you must have set up the original series recording after selecting the program in the Guide at the exact time of a first run showing (11pm, in your case), and not on a repeat entry in the Guide.  Then, even it The Daily Show is tagged as New for repeat showings, these will be ignored. 
    Third, another channel may air reruns of the program in your series recording, and the first showing of a rerun episode on the other channel may be tagged as "New".  These can be ignored in your series if you set the series option "Channel" to "Selected Channel".  Related to this, if there is both an SD and HD channel broadcasting you series program, you will record them both if the series option "Duplicates" is set to "Yes".  However, when the Channel option is set to "Selected Channel", the Duplicates Option is always effectively "No", regardless of what shows up on the options screen.  
    As for you missing two minutes,  I have sereral instances in which two programs start recording at the same time.  To the best of my recollection, whenever the warning message has appeared, ignoring it has not caused a loss of recording time.  You might have an older software version.  Newest is v.1.8.  Look at Menu--Settings--System Info.  Or, I might not have noticed the loss of minutes.  I regularly see up to a minute of previous programming at the start of a recording, or a few missing seconds at the beginning or end of a recording.  There are a lot of possibilities for that, but the DVR clock being incorrect is not one of them.  With RCN, the DVR clocks occasionally drifted off by as much as a minute and a half.

  • Pension issue Mid Month Leaving

    Dear All,
    As per rule sustem should deduct mid month joining/leaving/absences or transfer scenarios, the Pension/PF Basis will be correspondingly prorated. But our system is not doing this. In RT table i have found 3FC Pension Basis for Er c 01/2010                    0.00           6,500.00.
    Employee leaving date is 14.04.2010. system is picking pension amout as 541. Last year it was coming right.
    Please suggest.
    Ashwani

    Dear Jayanti,
    We required prorata basis pension in case of left employees and system is not doing this. This is the issue. As per our PF experts Pension amount should come on prorata basis for left employees in case they left mid of month.System is doing prorata basis last year but from this year it is deducting 541. I am giving two RT cases of different years.
    RT table for year 2010. DOL 26.04.2010
    /111 EPF Basis              01/2010                    0.00           8,750.00 
    /139 VPF Basis              01/2010                    0.00           8,750.00 
    /3F1 Ee PF contribution     01/2010                    0.00           1,050.00 
    /3F3 Er PF contribution     01/2010                    0.00             509.00 
    /3F5 Ee Mon PF contribution 01/2010                    0.00           1,050.00 
    /3F6 Ee Ann PF contribution 01/2010                    0.00          12,600.00 
    /3F9 PF adm chrgs * 1,00,00 01/2010                    0.00              96.25 
    /3FA PF basis for Ee contri 01/2010                    0.00           8,750.00 
    /3FB PF Basis for Er Contri 01/2010                    0.00           8,750.00 
    /3FJ VPF basis for Ee contr 01/2010                    0.00           8,750.00 
    /3FL PF Basis for Er Contri 01/2010                    0.00           6,500.00 
    /3F4 Er Pension contributio 01/2010                    0.00             541.00
    /3FC Pension Basis for Er c 01/2010                    0.00           6,500.00
    /3FB PF Basis for Er Contri 01/2010                    0.00           8,750.00
    /3FC Pension Basis for Er c 01/2010                    0.00           6,500.00
    /3FJ VPF basis for Ee contr 01/2010                    0.00           8,750.00
    /3FL PF Basis for Er Contri 01/2010                    0.00           6,500.00
    /3R3 Metro HRA Basis Amount 01/2010                    0.00           8,750.00
    1BAS Basic Salary           01/2010                    0.00           8,750.00
    RT table for year 2009. DOL 27.10.2009
                                                                                    /111 EPF Basis              07/2009                    0.00           9,016.13
    /139 VPF Basis              07/2009                    0.00           9,016.13
    /3F1 Ee PF contribution     07/2009                    0.00           1,082.00
    /3F3 Er PF contribution     07/2009                    0.00             628.00
    /3F5 Ee Mon PF contribution 07/2009                    0.00           1,082.00
    /3F6 Ee Ann PF contribution 07/2009                    0.00           8,822.00
    /3F9 PF adm chrgs * 1,00,00 07/2009                    0.00              99.18
    /3FA PF basis for Ee contri 07/2009                    0.00           9,016.00
    /3FB PF Basis for Er Contri 07/2009                    0.00           9,016.00
    /3FJ VPF basis for Ee contr 07/2009                    0.00           9,016.00
    /3FL PF Basis for Er Contri 07/2009                    0.00           5,452.00
    /3FB PF Basis for Er Contri 07/2009                    0.00           9,016.00 
    /3FC Pension Basis for Er c 07/2009                    0.00           5,452.00 
    /3FJ VPF basis for Ee contr 07/2009                    0.00           9,016.00 
    /3FL PF Basis for Er Contri 07/2009                    0.00           5,452.00 
    /3R4 Non-metro HRA Basis Am 07/2009                    0.00           9,016.13 
    1BAS Basic Salary           07/2009                    0.00           9,016.13 
    Now please suggest what to do. where is the problem  ? If have also checked EXIT_HINCALC0_002 but nothing written in it.
    With Regards
    Ashwani

  • Open PO Analysis - BW report issue

    Hello Friends
    I constructed a query in BW in order to show Open Purchase Orders. We have custom DSO populated with standard
    datasource 2lis_02_itm (Purcahse Order Item). In this DSO we mapped the field ELIKZ to the infoobject 0COMP_DEL
    (Delivery completed).
    We loaded the data from ECC system for all POs and found the following issue for Stock Transport Purchase orders (DocType = UB).
    We have a PO with 4 line items. For line items 10 and 20, Goods issued, Goods received and both the flags "Delivery
    complete" and "Final delivery" checked. For line items 30 and 40, only delivery indicator note is issued for zero
    quantity and Delivery complete flag is checked (Final delivery flag is not checked) in ECC system. For this PO, the
    delivery completion indicator is not properly updated in the DSO for line items 30 and 40. The data looks like the
    following:
    DOC_NUM     DOC_ITEM       DOCTYPE     COMP_DEL
    650000001       10     UB        X
    650000001       20     UB        X
    650000001       30     UB
    650000001       40     UB      
    When we run the Open PO analysis report on BW side this PO is appearing in the report but the same is closed in ECC
    system.
    Any help is appreciated in this regard.
    Thanks and Regards
    sampath

    Hi Priya and Reddy
       Thanks for your response.
                         Yes the indicator is checked in EKPO table for items 30 and 40 and delta is running regularly for more than 1 year and no issues with other POs. This is happening only for few POs of type Stock Transport (UB).
                        I already checked the changes in ME23N and the Delivery completed indicator was changed and it reflected in EKPO table. Further, i checked the PSA records for this PO and i am getting the records with the Delivery completed flag but when i update from PSA to DSO the delivery completed indicator is not updating properly.
                       In PSA, for item 30 i have the following entries. Record number 42 is capturing the value X for ELIKZ but after that i am getting two more records 43 and 44 with process key 10 and without X for ELIKZ. I think this is causing the problem.
    Record No.    Doc.No.                    Item              Processkey         Rocancel     Elikz
        41               6500000001            30                    11                            X           ---    
        42               6500000001            30                    11                            ---           X
        43               6500000001            30                    10                            X           ---
        44               6500000001            30                    10                            ---         ---
    (Here --- means blank)        
    Thanks and Regards
    sampath

  • HP LaserJet Enterprise 600 M602 driver issue

    Hello,
    I've got issue with 600-series printers. We use the latest UPD drivrer ver. 61.175.1.18849 and print from XenApp 6.5. The error occurs every time when users try to print jpg files from XenApp session. It only happens with 600 series printers and UPD.
    Also I've tried to assign native 600-series driver ver. 6.3.9600.16384 and it works good. But with that driver system says that it's color printer and it brokes our printing reports. These reports are very important for us. So we can't use printer and that driver as well.
    Printer installed on Windows Server 2012 R2. All clients are Windows 7 x64. XenApp Servers are Server 2008R2.
    Is it possible to get fixed UPD driver or correct native driver for Server 2012 R2?
    Regards,
    Anatoly

    I am sorry, but to get your issue more exposure I would suggest posting it in the commercial forums since this is a commercial printer. You can do this at Printers - LaserJet.
    Click on New Post.
    I hope this helps.
    Please click “Accept as Solution ” if you feel my post solved your issue, it will help others find the solution.
    Click the “Kudos Thumbs Up" on the right to say “Thanks” for helping!
    Gemini02
    I work on behalf of HP

Maybe you are looking for