Unable to set Conversation ID in UDF

I am trying to set Conversation ID in a Mapping UDF on a response message from ECC To XI to JMS Receiver.
GlobalContainer globalContainer ;
globalContainer = container.getGlobalContainer();
Map param = globalContainer.getParameters();
param.put(StreamTransformationConstants.CONVERSATION_ID, "1234ConversationID);
The mapping executes without any error. But for some reason, the conversation Id does not show up under the SOAP:Main (SAP:ConversationId element) header at all.
Any ideas?
Thanks,
Badari

Hi Badari/Aamir,
Both are same. Have a look at the standard SAP java program for StreamTransformation Constants.
package com.sap.aii.mapping.api;
public interface StreamTransformationConstants
    public static final String MESSAGE_CLASS = "MessageClass";
    public static final String VERSION_MAJOR = "VersionMajor";
    public static final String VERSION_MINOR = "VersionMinor";
    public static final String PROCESSING_MODE = "ProcessingMode";
    public static final String MESSAGE_ID = "MessageId";
    public static final String REF_TO_MESSAGE_ID = "RefToMessageId";
    public static final String CONVERSATION_ID = "ConversationId";
    public static final String TIME_SENT = "TimeSent";
    public static final String SENDER_PARTY = "SenderParty";
    public static final String SENDER_PARTY_AGENCY = "SenderPartyAgency";
    public static final String SENDER_PARTY_SCHEME = "SenderPartyScheme";
    public static final String SENDER_SERVICE = "SenderService";
    public static final String RECEIVER_PARTY = "ReceiverParty";
    public static final String RECEIVER_PARTY_AGENCY = "ReceiverPartyAgency";
    public static final String RECEIVER_PARTY_SCHEME = "ReceiverPartyScheme";
    public static final String RECEIVER_SERVICE = "ReceiverService";
    public static final String INTERFACE = "Interface";
    public static final String INTERFACE_NAMESPACE = "InterfaceNamespace";
    public static final String MAPPING_TRACE = "MappingTrace";
    public static final String CONTENT_TYPE = "ContentType";
    public static final String DYNAMIC_CONFIGURATION = "DynamicConfiguration";
@deprecated Field SENDER_NAME is deprecated
    public static final String SENDER_NAME = "SenderName";
@deprecated Field SENDER_NAMESPACE is deprecated
    public static final String SENDER_NAMESPACE = "SenderNamespace";
@deprecated Field SENDER_SYSTEM is deprecated
    public static final String SENDER_SYSTEM = "SenderSystem";
    public static final String RECEIVER_NAME = "ReceiverName";
    public static final String RECEIVER_NAMESPACE = "ReceiverNamespace";
@deprecated Field RECEIVER_SYSTEM is deprecated
    public static final String RECEIVER_SYSTEM = "ReceiverSystem";
@deprecated Field DIRECTION is deprecated
    public static final String DIRECTION = "Direction";
@deprecated Field DIRECTION_REQUEST is deprecated
    public static final String DIRECTION_REQUEST = "Request";
@deprecated Field DIRECTION_RESPONSE is deprecated
    public static final String DIRECTION_RESPONSE = "Response";
@deprecated Field DIRECTION_FAULT is deprecated
    public static final String DIRECTION_FAULT = "Fault";
Thanks,
Gujjeti

Similar Messages

  • The ADO NET Source was unable to process the data. ORA-64203: Destination buffer too small to hold CLOB data after character set conversion.

     We developed a SSIS Package to pull the data From Oracle source to Sql Server 2012. Here we used ADO.Net source to pull the records from Source but getting the below error after pulling some 40K records.
      [ADO NET Source [2]] Error: The ADO NET Source was unable to process the data. ORA-64203: Destination buffer too small to hold CLOB data after character set conversion.
    [SSIS.Pipeline] Error: SSIS Error Code DTS_E_PRIMEOUTPUTFAILED. 
     The PrimeOutput method on ADO NET Source returned error code 0xC02090F5. 
     The component returned a failure code when the pipeline engine called PrimeOutput(). 
    The meaning of the failure code is defined by the component, 
    but the error is fatal and the pipeline stopped executing. 
     There may be error messages posted before this with more 
    information about the failure.
    Anything that we can do to fix this?

    Hi,
      Tried both....
      * Having schema type as Nvarchar(max). - Getting the same error.
      * Instead of ADO.Net Source used OLEDB Source with driver as " Oracle Provide for OLE DB" Getting error as below.
           [OLE DB Source [478]] Error: SSIS Error Code DTS_E_OLEDBERROR.  An OLE DB error has occurred. Error code: 0x80004005.
           [SSIS.Pipeline] Error: SSIS Error Code DTS_E_PRIMEOUTPUTFAILED.  The PrimeOutput method on OLE DB Source returned error code 0xC0202009.  The component returned a failure
    code when the pipeline engine called PrimeOutput(). The meaning of the failure code is defined by the component, but the error is fatal and the pipeline stopped executing.  There may be error messages posted before this with more information about the
    failure.
    Additional Info:
       * Here the Source task is getting failed not the conversion or destination task.
    Thanks,
    Loganathan A.

  • Fixing a US7ASCII - WE8ISO8859P1 Character Set Conversion Disaster

    In hopes that it might be helpful in the future, here's the procedure I followed to fix  a disastrous unintentional US7ASCII on 9i to WE8ISO8859P1 on 10g migration.
    BACKGROUND
    Oracle has multiple character sets, ranging from US7ASCII to AL32UTF16.
    US7ASCII, of course, is a cheerful 7 bit character set, holding the basic ASCII characters sufficient for the English language.
    However, it also has a handy feature: character fields under US7ASCII will accept characters with values > 128. If you have a web application, users can type (or paste) Us with umlauts, As with macrons, and quite a few other funny-looking characters.
    These will be inserted into the database, and then -- if appropriately supported -- can be selected and displayed by your app.
    The problem is that while these characters can be present in a VARCHAR2 or CLOB column, they are not actually legal. If you try within Oracle to convert from US7ASCII to WE8ISO8859P1 or any other character set, Oracle recognizes that these characters with values greater than 127 are not valid, and will replace them with a default "unknown" character. In the case of a change from US7ASCII to WE8ISO8859P1, it will change them to 191, the upside down question mark.
    Oracle has a native utility, introduced in 8i, called csscan, which assists in migrating to different character sets. This has been replaced in newer versions with the Database MIgration Assistant for Unicode (DMU), which is the new recommended tool for 11.2.0.3+.
    These tools, however, do no good unless they are run. For my particular client, the operations team took a database running 9i and upgraded it to 10g, and as part of that process the character set was changed from US7ASCII to WE8ISO8859P1. The database had a large number of special characters inserted into it, and all of these abruptly turned into upside-down question marks. The users of the application didn't realize there was a problem until several weeks later, by which time they had put a lot of new data into the system. Rollback was not possible.
    FIXING THE PROBLEM
    How fixable this problem is and the acceptable methods which can be used depend on the application running on top of the database. Fortunately, the client app was amenable.
    (As an aside note: this approach does not use csscan -- I had done something similar previously on a very old system and decided it would take less time in this situation to revamp my old procedures and not bring a new utility into the mix.)
    We will need to separate approaches -- one to fix the VARCHAR2 & CHAR fields,  and a second for CLOBs.
    In order to set things up, we created two environments. The first was a clone of production as it is now, and the second a clone from before the upgrade & character set change. We will call these environments PRODCLONE and RESTORECLONE.
    Next, we created a database link, OLD6. This allows PRODCLONE to directly access RESTORECLONE. Since they were cloned with the same SID, establishing the link needed the global_names parameter set to false.
    alter system set global_names=false scope=memory;
    CREATE PUBLIC DATABASE LINK OLD6
    CONNECT TO DBUSERNAME
    IDENTIFIED BY dbuserpass
    USING 'restoreclone:1521/MYSID';
    Testing the link...
    SQL> select count(1) from users@old6;
      COUNT(1)
           454
    Here is a row in a table which contains illegal characters. We are accessing RESTORECLONE from PRODCLONE via our link.
    PRODCLONE> select dump(title) from my_contents@old6 where pk1=117286;
    DUMP(TITLE)
    Typ=1 Len=49: 78,67,76,69,88,45,80,78,174,32,69,120,97,109,32,83,116,121,108,101
    ,32,73,110,116,101,114,97,99,116,105,118,101,32,82,101,118,105,101,119,32,81,117
    ,101,115,116,105,111,110,115
    By comparison, a dump of that row on PRODCLONE's my_contents gives:
    PRODCLONE> select dump(title) from my_contents where pk1=117286;
    DUMP(TITLE)
    Typ=1 Len=49: 78,67,76,69,88,45,80,78,191,32,69,120,97,109,32,83,116,121,108,101
    ,32,73,110,116,101,114,97,99,116,105,118,101,32,82,101,118,105,101,119,32,81,117
    ,101,115,116,105,111,110,115
    Note that the "174" on RESTORECLONE was changed to "191" on PRODCLONE.
    We can manually insert CHR(174) into our PRODCLONE and have it display successfully in the application.
    However, I tried a number of methods to copy the data from RESTORECLONE to PRODCLONE through the link, but entirely without success. Oracle would recognize the character as invalid and silently transform it.
    Eventually, I located a clever workaround at this link:
    https://kr.forums.oracle.com/forums/thread.jspa?threadID=231927
    It works like this:
    On RESTORECLONE you create a view, vv, with UTL_RAW:
    RESTORECLONE> create or replace view vv as select pk1,utl_raw.cast_to_raw(title) as title from my_contents;
    View created.
    This turns the title to raw on the RESTORECLONE.
    You can now convert from RAW to VARCHAR2 on the PRODCLONE database:
    PRODCLONE> select dump(utl_raw.cast_to_varchar2 (title)) from vv@old6 where pk1=117286;
    DUMP(UTL_RAW.CAST_TO_VARCHAR2(TITLE))
    Typ=1 Len=49: 78,67,76,69,88,45,80,78,174,32,69,120,97,109,32,83,116,121,108,101
    ,32,73,110,116,101,114,97,99,116,105,118,101,32,82,101,118,105,101,119,32,81,117
    ,101,115,116,105,111,110,115
    The above works because oracle on PRODCLONE never knew that our TITLE string on RESTORE was originally in  US7ASCII, so it was unable to do its transparent character set conversion.
    PRODCLONE> update my_contents set title=( select utl_raw.cast_to_varchar2 (title) from vv@old6 where pk1=117286) where pk1=117286;
    PRODCLONE> select dump(title) from my_contents where pk1=117286;
    DUMP(UTL_RAW.CAST_TO_VARCHAR2(TITLE))
    Typ=1 Len=49: 78,67,76,69,88,45,80,78,174,32,69,120,97,109,32,83,116,121,108,101
    ,32,73,110,116,101,114,97,99,116,105,118,101,32,82,101,118,105,101,119,32,81,117
    ,101,115,116,105,111,110,115
    Excellent! The "174" character has survived the transfer and is now in place on PRODCLONE.
    Now that we have a method to move the data over, we have to identify which columns /tables have character data that was damaged by the conversion. We decided we could ignore anything with a length smaller than 10 -- such fields in our application would be unlikely to have data with invalid characters.
    RESTORECLONE> select count(1) from user_tab_columns where data_type in ('CHAR','VARCHAR2') and data_length > 10;
       COUNT(1)
        533
    By converting a field to WE8ISO8859P1, and then comparing it with the original, we can see if the characters change:
    RESTORECLONE> select count(1) from my_contents where title != convert (title,'WE8ISO8859P1','US7ASCII') ;
      COUNT(1)
         10568
    So 10568 rows have characters which were transformed  into 191s as part of the original conversion.
    [ As an aside, we can't use CONVERT() on LOBs -- for them we will need another approach, outlined further below.
    RESTOREDB> select count(1) from my_contents where main_data != convert (convert(main_DATA,'WE8ISO8859P1','US7ASCII'),'US7ASCII','WE8ISO8859P1') ;
    select count(1) from my_contents where main_data != convert (convert(main_DATA,'WE8ISO8859P1','US7ASCII'),'US7ASCII','WE8ISO8859P1')
    ERROR at line 1:
    ORA-00932: inconsistent datatypes: expected - got CLOB
    Anyway, now that we can identify VARCHAR2 fields which need to be checked, we can put together a PL/SQL stored procedure to do it for us:
    create or replace procedure find_us7_strings
    (table_name varchar2,
    fix_col varchar2 )
    authid current_user
    as
    orig_sql varchar2(1000);
    begin
    orig_sql:='insert into cnv_us7(mytablename,myindx,mycolumnname)  select '''||table_name||''',pk1,'''||fix_col||''' from '||table_name||' where '||fix_col||' !=  CONVERT(CONVERT('||fix_col||',''WE8ISO8859P1''),''US7ASCII'') and '||fix_col||' is not null';
    -- Uncomment if debugging:
    -- dbms_output.put_line(orig_sql);
      execute immediate orig_sql;
    end;
    And create a table to store the information as to which tables, columns, and rows have the bad characters:
    drop table cnv_us7;
    create table cnv_us7 (mytablename varchar2(50), myindx number,      mycolumnname varchar2(50) ) tablespace myuser_data;
    create index list_tablename_idx on cnv_us7(mytablename) tablespace myuser_indx;
    With a SQL-generating SQL script, we can iterate through all the tables/columns we want to check:
    --example of using the data: select title from my_contents where pk1 in (select myindx from cnv_us7)
    set head off pagesize 1000 linesize 120
    spool runme.sql
    select 'exec find_us7_strings ('''||table_name||''','''||column_name||'''); ' from user_tab_columns
          where
              data_type in ('CHAR','VARCHAR2')
              and table_name in (select table_name from user_tab_columns where column_name='PK1' and  table_name not  in ('HUGETABLEIWANTTOEXCLUDE','ANOTHERTABLE'))
              and char_length > 10
              order by table_name,column_name;
    spool off;
    set echo on time on timing on feedb on serveroutput on;
    spool output_of_runme
    @./runme.sql
    spool off;
    Which eventually gives us the following inserted into CNV_US7:
    20:48:21 SQL> select count(1),mycolumnname,mytablename from cnv_us7 group by mytablename,mycolumnname;
             4 DESCRIPTION                                        MY_FORUMS
         21136 TITLE                                              MY_CONTENTS
    Out of 533 VARCHAR2s and CHARs, we only had five or six columns that needed fixing
    We create our views on  RESTOREDB:
    create or replace view my_forums_vv as select pk1,utl_raw.cast_to_raw(description) as description from forum_main;
    create or replace view my_contents_vv as select pk1,utl_raw.cast_to_raw(title) as title from my_contents;
    And then we can fix it directly via sql:
    update my_contents taborig1 set TITLE= (select utl_raw.cast_to_varchar2 (TITLE) from my_contents_vv@old6 where pk1=taborig1.pk1)
    where pk1 in (
    select tabnew.pk1 from my_contents@old6 taborig,my_contents tabnew,cnv_us7@old6
          where taborig.pk1=tabnew.pk1
              and myindx=tabnew.pk1
              and mycolumnname='TITLE'
              and mytablename='MY_CONTENTS'
              and convert(taborig.TITLE,'US7ASCII','WE8ISO8859P1') = tabnew.TITLE );
    Note this part:
          "and convert(taborig.TITLE,'US7ASCII','WE8ISO8859P1') = tabnew.TITLE "
    This checks to verify that the TITLE field on the PRODCLONE and RESTORECLONE are the same (barring character set issues). This is there  because if the users have changed TITLE  -- or any other field -- on their own between the time of the upgrade and now, we do not want to overwrite their changes. We make the assumption that as part of the process, they may have changed the bad character on their own.
    We can also create a stored procedure which will execute the SQL for us:
    create or replace procedure fix_us7_strings
    (TABLE_NAME varchar2,
    FIX_COL varchar2 )
    authid current_user
    as
    orig_sql varchar2(1000);
    TYPE cv_type IS REF CURSOR;
    orig_cur cv_type;
    begin
    orig_sql:='update '||TABLE_NAME||' taborig1 set '||FIX_COL||'= (select utl_raw.cast_to_varchar2 ('||FIX_COL||') from '||TABLE_NAME||'_vv@old6 where pk1=taborig1.pk1)
    where pk1 in (
    select tabnew.pk1 from '||TABLE_NAME||'@old6 taborig,'||TABLE_NAME||' tabnew,cnv_us7@old6
          where taborig.pk1=tabnew.pk1
              and myindx=tabnew.pk1
              and mycolumnname='''||FIX_COL||'''
              and mytablename='''||TABLE_NAME||'''
              and convert(taborig.'||FIX_COL||',''US7ASCII'',''WE8ISO8859P1'') = tabnew.'||FIX_COL||')';
    dbms_output.put_line(orig_sql);
    execute immediate orig_sql;
    end;
    exec fix_us7_strings('MY_FORUMS','DESCRIPTION');
    exec fix_us7_strings('MY_CONTENTS','TITLE');
    commit;
    To validate this before and after, we can run something like:
    select dump(description) from my_forums where pk1 in (select myindx from cnv_us7@old6 where mytablename='MY_FORUMS');
    The above process fixes all the VARCHAR2s and CHARs. Now what about the CLOB columns?
    Note that we're going to have some extra difficulty here, not just because we are dealing with CLOBs, but because we are working with CLOBs in 9i, whose functions have less CLOB-related functionality.
    This procedure finds invalid US7ASCII strings inside a CLOB in 9i:
    create or replace procedure find_us7_clob
    (table_name varchar2,
    fix_col varchar2)
    authid current_user
    as
      orig_sql varchar2(1000);
      type cv_type is REF CURSOR;
      orig_table_cur cv_type;
      my_chars_read NUMBER;
      my_offset NUMBER;
      my_problem NUMBER;
      my_lob_size NUMBER;
      my_indx_var NUMBER;
      my_total_chars_read NUMBER;
      my_output_chunk VARCHAR2(4000);
      my_problem_flag NUMBER;
      my_clob CLOB;
      my_total_problems NUMBER;
      ins_sql VARCHAR2(4000);
    BEGIN
       DBMS_OUTPUT.ENABLE(1000000);
       orig_sql:='select pk1,dbms_lob.getlength('||FIX_COL||') as cloblength,'||fix_col||' from '||table_name||' where dbms_lob.getlength('||fix_col||') >0 and '||fix_col||' is not null order by pk1';
       open orig_table_cur for orig_sql;
       my_total_problems := 0;
       LOOP
            FETCH orig_table_cur INTO my_indx_var,my_lob_size,my_clob;
                    EXIT WHEN orig_table_cur%NOTFOUND;
            my_offset :=1;
            my_chars_read := 512;
            my_problem_flag :=0;
            WHILE my_offset < my_lob_size and my_problem_flag =0
                    LOOP
                    DBMS_LOB.READ(my_clob,my_chars_read,my_offset,my_output_chunk);
                    my_offset := my_offset + my_chars_read;
                    IF my_output_chunk != CONVERT(CONVERT(my_output_chunk,'WE8ISO8859P1'),'US7ASCII')
                            THEN
                            -- DBMS_OUTPUT.PUT_LINE('Problem with '||my_indx_var);
                            -- DBMS_OUTPUT.PUT_LINE(my_output_chunk);
                            my_problem_flag:=1;
                    END IF;
            END LOOP;
            IF my_problem_flag=1
                    THEN my_total_problems := my_total_problems +1;
                    ins_sql:='insert into cnv_us7(mytablename,myindx,mycolumnname) values ('''||table_name||''','||my_indx_var||','''||fix_col||''')';
                    execute immediate ins_sql;
                    END IF;
       END LOOP;
       DBMS_OUTPUT.PUT_LINE('We found '||my_total_problems||' problem rows in table '||table_name||', column '||fix_col||'.');
    END;
    And we can use SQL-generating SQL to find out which CLOBs have issues, out of all the ones in the database:
    RESTOREDB> select 'exec find_us7_clob('''||table_name||''','''||column_name||''');' from user_tab_columns where data_type='CLOB';
    exec find_us7_clob('MY_CONTENTS','DATA');
    After completion, the CNV_US7 table looked like this:
    RESTOREDB> set linesize 120 pagesize 100;
    RESTOREDB>  select count(1),mytablename,mycolumnname from cnv_us7
       where mytablename||' '||mycolumnname in (select table_name||' '||column_name from user_tab_columns
             where data_type='CLOB' )
          group by mytablename,mycolumnname;
      COUNT(1) MYTABLENAME                                        MYCOLUMNNAME
         69703 MY_CONTENTS                                  DATA
    On RESTOREDB, our 9i version, we will use this procedure (found many years ago on the internet):
    create or replace procedure CLOB2BLOB (p_clob in out nocopy clob, p_blob in out nocopy blob) is
    -- transforming CLOB to BLOB
    l_off number default 1;
    l_amt number default 4096;
    l_offWrite number default 1;
    l_amtWrite number;
    l_str varchar2(4096 char);
    begin
    loop
    dbms_lob.read ( p_clob, l_amt, l_off, l_str );
    l_amtWrite := utl_raw.length ( utl_raw.cast_to_raw( l_str) );
    dbms_lob.write( p_blob, l_amtWrite, l_offWrite,
    utl_raw.cast_to_raw( l_str ) );
    l_offWrite := l_offWrite + l_amtWrite;
    l_off := l_off + l_amt;
    l_amt := 4096;
    end loop;
    exception
    when no_data_found then
    NULL;
    end;
    We can test out the transformation of CLOBs to BLOBs with a single row like this:
    drop table my_contents_lob;
    Create table my_contents_lob (pk1 number,data blob);
    DECLARE
          v_clob CLOB;
          v_blob BLOB;
        BEGIN
          SELECT data INTO v_clob FROM my_contents WHERE pk1 = 16 ;
          INSERT INTO my_contents_lob (pk1,data) VALUES (16,empty_blob() );
          SELECT data INTO v_blob FROM my_contents_lob WHERE pk1=16 FOR UPDATE;
          clob2blob (v_clob, v_blob);
        END;
    select dbms_lob.getlength(data) from my_contents_lob;
    DBMS_LOB.GETLENGTH(DATA)
                                 329
    SQL> select utl_raw.cast_to_varchar2(data) from my_contents_lob;
    UTL_RAW.CAST_TO_VARCHAR2(DATA)
    Lorem ipsum dolor sit amet, consectetur adipisicing elit, sed do eiusmod tempor incididunt ut labore et dolore magna aliqua. Ut enim ad minim veniam...
    Now we need to push it through a loop. Unfortunately, I had trouble making the "SELECT INTO" dynamic. Thus I used a version of the procedure for each table. It's aesthetically displeasing, but at least it worked.
    create table my_contents_lob(pk1 number,data blob);
    create index my_contents_lob_pk1 on my_contents_lob(pk1) tablespace my_user_indx;
    create or replace procedure blob_conversion_my_contents
    (table_name varchar2,
    fix_col varchar2)
    authid current_user
    as
      orig_sql varchar2(1000);
      type cv_type is REF CURSOR;
      orig_table_cur cv_type;
      my_chars_read NUMBER;
      my_offset NUMBER;
      my_problem NUMBER;
      my_lob_size NUMBER;
      my_indx_var NUMBER;
      my_total_chars_read NUMBER;
      my_output_chunk VARCHAR2(4000);
      my_problem_flag NUMBER;
      my_clob CLOB;
      my_blob BLOB;
      my_total_problems NUMBER;
      new_sql VARCHAR2(4000);
    BEGIN
      DBMS_OUTPUT.ENABLE(1000000);
       orig_sql:='select pk1,dbms_lob.getlength('||FIX_COL||') as cloblength,'||fix_col||' from '||table_name||' where pk1 in (select myindx from cnv_us7 where mytablename='''||TABLE_NAME||''' and mycolumnname='''||FIX_COL||''') order by pk1';
       open orig_table_cur for orig_sql;
       LOOP
            FETCH orig_table_cur INTO my_indx_var,my_lob_size,my_clob;
                    EXIT WHEN orig_table_cur%NOTFOUND;
            new_sql:='INSERT INTO '||table_name||'_lob(pk1,'||fix_col||') values ('||my_indx_var||',empty_blob() )';
            dbms_output.put_line(new_sql);
          execute immediate new_sql;
    -- Here's the bit that I had trouble making dynamic. Feel free to let me know what I am doing wrong.
    -- new_sql:='SELECT '||fix_col||' INTO my_blob from '||table_name||'_lob where pk1='||my_indx_var||' FOR UPDATE';
    --        dbms_output.put_line(new_sql);
            select data into my_blob from my_contents_lob where pk1=my_indx_var FOR UPDATE;
          clob2blob(my_clob,my_blob);
       END LOOP;
       CLOSE orig_table_cur;
      DBMS_OUTPUT.PUT_LINE('Completed program');
    END;
    exec blob_conversion_my_contents('MY_CONTENTS','DATA');
    Verify that things work properly:
    select dump( utl_raw.cast_to_varchar2(data))  from my_contents_lob where pk1=xxxx;
    This should let you see see characters > 150. Thus, the method works.
    We can now take this data, export it from RESTORECLONE
    exp file=a.dmp buffer=4000000 userid=system/XXXXXX tables=my_user.my_contents rows=y
    and import the data on prodclone
    imp file=a.dmp fromuser=my_user touser=my_user userid=system/XXXXXX buffer=4000000;
    For paranoia's sake, double check that it worked properly:
    select dump( utl_raw.cast_to_varchar2(data))  from my_contents_lob;
    On our 10g PRODCLONE, we'll use these stored procedures:
    CREATE OR REPLACE FUNCTION CLOB2BLOB(L_CLOB CLOB) RETURN BLOB IS
    L_BLOB BLOB;
    L_SRC_OFFSET NUMBER;
    L_DEST_OFFSET NUMBER;
    L_BLOB_CSID NUMBER := DBMS_LOB.DEFAULT_CSID;
    V_LANG_CONTEXT NUMBER := DBMS_LOB.DEFAULT_LANG_CTX;
    L_WARNING NUMBER;
    L_AMOUNT NUMBER;
    BEGIN
    DBMS_LOB.CREATETEMPORARY(L_BLOB, TRUE);
    L_SRC_OFFSET := 1;
    L_DEST_OFFSET := 1;
    L_AMOUNT := DBMS_LOB.GETLENGTH(L_CLOB);
    DBMS_LOB.CONVERTTOBLOB(L_BLOB,
    L_CLOB,
    L_AMOUNT,
    L_SRC_OFFSET,
    L_DEST_OFFSET,
    1,
    V_LANG_CONTEXT,
    L_WARNING);
    RETURN L_BLOB;
    END;
    CREATE OR REPLACE FUNCTION BLOB2CLOB(L_BLOB BLOB) RETURN CLOB IS
    L_CLOB CLOB;
    L_SRC_OFFSET NUMBER;
    L_DEST_OFFSET NUMBER;
    L_BLOB_CSID NUMBER := DBMS_LOB.DEFAULT_CSID;
    V_LANG_CONTEXT NUMBER := DBMS_LOB.DEFAULT_LANG_CTX;
    L_WARNING NUMBER;
    L_AMOUNT NUMBER;
    BEGIN
    DBMS_LOB.CREATETEMPORARY(L_CLOB, TRUE);
    L_SRC_OFFSET := 1;
    L_DEST_OFFSET := 1;
    L_AMOUNT := DBMS_LOB.GETLENGTH(L_BLOB);
    DBMS_LOB.CONVERTTOCLOB(L_CLOB,
    L_BLOB,
    L_AMOUNT,
    L_SRC_OFFSET,
    L_DEST_OFFSET,
    1,
    V_LANG_CONTEXT,
    L_WARNING);
    RETURN L_CLOB;
    END;
    And now, for the piece de' resistance, we need a BLOB to CLOB conversion that assumes that the BLOB data is stored initially in WE8ISO8859P1.
    To find correct CSID for WE8ISO8859P1, we can use this query:
    select nls_charset_id('WE8ISO8859P1') from dual;
    Gives "31"
    create or replace FUNCTION BLOB2CLOBASC(L_BLOB BLOB) RETURN CLOB IS
    L_CLOB CLOB;
    L_SRC_OFFSET NUMBER;
    L_DEST_OFFSET NUMBER;
    L_BLOB_CSID NUMBER := 31;      -- treat blob as  WE8ISO8859P1
    V_LANG_CONTEXT NUMBER := 31;   -- treat resulting clob as  WE8ISO8850P1
    L_WARNING NUMBER;
    L_AMOUNT NUMBER;
    BEGIN
    DBMS_LOB.CREATETEMPORARY(L_CLOB, TRUE);
    L_SRC_OFFSET := 1;
    L_DEST_OFFSET := 1;
    L_AMOUNT := DBMS_LOB.GETLENGTH(L_BLOB);
    DBMS_LOB.CONVERTTOCLOB(L_CLOB,
    L_BLOB,
    L_AMOUNT,
    L_SRC_OFFSET,
    L_DEST_OFFSET,
    L_BLOB_CSID,
    V_LANG_CONTEXT,
    L_WARNING);
    RETURN L_CLOB;
    END;
    select dump(dbms_lob.substr(blob2clobasc(data),4000,1)) from my_contents_lob;
    Now, we can compare these:
    select dbms_lob.compare(blob2clob(old.data),new.data) from  my_contents new,my_contents_lob old where new.pk1=old.pk1;
    DBMS_LOB.COMPARE(BLOB2CLOB(OLD.DATA),NEW.DATA)
                                                                 0
                                                                 0
                                                                 0
    Vs
    select dbms_lob.compare(blob2clobasc(old.data),new.data) from  my_contents new,my_contents_lob old where new.pk1=old.pk1;
    DBMS_LOB.COMPARE(BLOB2CLOBASC(OLD.DATA),NEW.DATA)
                                                                   -1
                                                                   -1
                                                                   -1
    update my_contents a set data=(select blob2clobasc(data) from my_contents_lob b where a.pk1= b.pk1)
        where pk1 in (select al.pk1 from my_contents_lob al where dbms_lob.compare(blob2clob(al.data),a.data) =0 );
    SQL> select dump(dbms_lob.substr(data,4000,1)) from my_contents where pk1 in (select pk1 from my_contents_lob);
    Confirms that we're now working properly.
    To run across all the _LOB tables we've created:
    [oracle@RESTORECLONE ~]$ exp file=all_fixed_lobs.dmp buffer=4000000 userid=my_user/mypass tables=MY_CONTENTS_LOB,MY_FORUM_LOB...
    [oracle@RESTORECLONE ~]$ scp all_fixed_lobs.dmp jboulier@PRODCLONE:/tmp
    And then on PRODCLONE we can import:
    imp file=all_fixed_lobs.dmp buffer=4000000 userid=system/XXXXXXX fromuser=my_user touser=my_user
    Instead of running the above update statement for all the affected tables, we can use a simple stored procedure:
    create or replace procedure fix_us7_CLOBS
      (TABLE_NAME varchar2,
         FIX_COL varchar2 )
        authid current_user
        as
         orig_sql varchar2(1000);
         bak_sql  varchar2(1000);
        begin
        dbms_output.put_line('Creating '||TABLE_NAME||'_PRECONV to preserve the original data in the table');
        bak_sql:='create table '||TABLE_NAME||'_preconv as select pk1,'||FIX_COL||' from '||TABLE_NAME||' where pk1 in (select pk1 from '||TABLE_NAME||'_LOB) ';
        execute immediate bak_sql;
        orig_sql:='update '||TABLE_NAME||' tabnew set '||FIX_COL||'= (select blob2clobasc ('||FIX_COL||') from '||TABLE_NAME||'_LOB taborig where tabnew.pk1=taborig.pk1)
       where pk1 in (
       select a.pk1 from '||TABLE_NAME||'_LOB a,'||TABLE_NAME||' b
          where a.pk1=b.pk1
                 and dbms_lob.compare(blob2clob(a.'||FIX_COL||'),b.'||FIX_COL||') = 0 )';
        -- dbms_output.put_line(orig_sql);
        execute immediate orig_sql;
       end;
    Now we can run the procedure and it fixes everything for our previously-broken tables, keeping the changed rows -- just in case -- in a table called table_name_PRECONV.
    set serveroutput on time on timing on;
    exec fix_us7_clobs('MY_CONTENTS','DATA');
    commit;
    After confirming with the client that the changes work -- and haven't noticeably broken anything else -- the same routines can be carefully run against the actual production database.

    We converted using the database using scripts I developed. I'm not quite sure how we converted is relevant, other than saying that we did not use the Oracle conversion utility (not csscan, but the GUI Java tool).
    A summary:
    1) We replaced the lossy characters by parsing a csscan output file
    2) After re-scanning with csscan and coming up clean, our DBA converted the database to AL32UTF8 (changed the parameter file, changing the character set, switched the semantics to char, etc).
    3) Final step was changing existing tables to use char semantics by changing the table schema for VARCHAR2 columns
    Any specific steps I cannot easily answer, I worked with a DBA at our company to do this work. I handled the character replacement / DDL changes and the DBA ran csscan & performed the database config changes.
    Our actual error message:
    ORA-31011: XML parsing failed
    ORA-19202: Error occurred in XML processing
    LPX-00210: expected '<' instead of '�Error at line 1
    31011. 00000 - "XML parsing failed"
    *Cause:    XML parser returned an error while trying to parse the document.
    *Action:   Check if the document to be parsed is valid.
    Error at Line: 24 Column: 15
    This seems to match the the document ID referenced below. I will ask our DBA to pull it up and review it.
    Please advise if more information is needed from my end.

  • Unable to set node credentials...

    I have setup the server as an OD master, and up until now it hasn't had problems binding, although I haven't been able to get any users to authenticate for any services, other than my local admin account.
    I haven't made any changes, but now when opening ServerPrefs, I get the following error:
    "Unable to set node credentials for /LDAPv3/127.0.0.1 with the record name server.mydomain.com$ There may be a problem with the Open Directory service."
    Any ideas where I went wrong?

    It isn't something that a patch is going to cure.
    I believe it does have something to do with the network and probably therefore the ip address / naming as the authentication does seem to be tied to the network location.
    I've managed to move things on a little in that I have bound user accounts on other machines such that when someone logs onto their laptop it will tell them they have an account on the server and provide a logon after which their account is bound and (I think) marked as 'managed' in their Preferences | Accounts. This provides additional sections in iCal and Address Book although these don't work. The only person who can log into the web stuff such as the wiki is the server admin and I still can't access the Server Preferences or authenticate the Server Admin so can't make much in the way of changes.
    I've submitted my application to Apple for support using the card provided with the Mac Mini. This was last Thursday and I've not yet received confirmation that this is live. As soon as it is I'll be having a long conversation with Apple Support.
    I sort of sympathise with Apple as the server is clearly at least an order of magnitude more complicated than the admin involved with just a desktop. Having said this, however, they do need to get this sort of thing sorted by making it slicker, providing more online support for these problems if they have any aspirations towards the smaller office or even home environment which don't have dedicated IT support staff.

  • Unable to  set some dates  in Adobe bridge cs2

    HI ALL
    Our keywording team using  the  Adobe bridge CS2.
    Kindly  hep me to understand Why  some dates  Like
    1/1/1970
    1/1/1642
    1/1/1890
    etc  .....
    I unable to  set in  ITPC image Tab  information .............
    Lakhan

    Hi
    Thanks for revert ......
    I  tried   with  CS2 & CS3  but only  specific  dates like (1/1/1970) i above  mentioned  is still  not   reflecting in IPTC image  info   ........
    Please suggest....

  • Illustrator won't run: "unable to set maximum number of files to be opened"

    I have an intel macbook pro (core duo with 2GB ram) and a copy of purchased adobe creative suite 2. Other applications in the suite like photoshop, acrobat, etc can run on my mac without any problem. But when I try to run illustrator, it just shows an error message "Unable to set maximum number of files to be opened" and quits. I also installed a trial version of illustrator cs3 and got the same problem. Can anyone help? Thanks.
    CC

    Hi, I realize this is an old thread but it was unanswered and the issue has come up again in a newer thread:
    http://forums.adobe.com/message/2534652
    If the original poster sees this, can you reply tin the new thread and let us know if you were ever to solve this problem?
    Thanks!

  • Service Desk Unable to set status to Customer Action and Proposed Solution

    Hi Everyone,
    We are unable to set a service desk issue to the status Customer Action or Proposed solution.
    I have someone create a new issue in the bakend system and when I receive the issue I assign it to myself as the processor. When I try to set teh status to Customer action it is grayed out.
    I have checked the following:
    1. As a test I assigned My user with sap_all profile so does not appear to be authroisation related
    2. The status profile has all six standard values (New, customer action, In Process, Sent to SAP, Proposed Solution, Confirmed).
    3. The transaction types have been defined using the standard profiles for Text Det. Procedure, Partnet Determination Procedure, Status Profile. The Partner Function ORG is set to Sold-To-Party , The date profile is a SRV_SLA_ITEM and the Action profile is set to SLFN0001_ADVANCED. The Org Data Prof is set to a customer defined one: ZSSC000000001.
    4. I have checked transaction ppoma_crm and the org structure and my user is assigned to a team lead position.
    Your assistance is greatly appreciated.
    Michael

    Hi Michael,
    please check your status profile for the "lowest status no." and "highest status no." values of the status values as Alex mentioned!
    You can find this in SPRO under "SAP Solution Manager> Scenario-Specific Settings>Service Desk>Service Desk>Status Profile".
    If the "high" value of the status "New" is lower than the status number of status "Customer Action" than you can not set the message from "New" to "Customer Action"!
    Hope this helps,
    Christoph

  • Unable to set the selected value for a ADF LOV

    Hi,
    I am developing an application using jdeveloper 10.1.3 and Struts.
    I have created a ADF LOV on my JSP Page as given below:
    <html:select property="lob" onchange="javascript:selectLOB(this.form)">
    <html:optionsCollection label="prompt" value="index" property="lob.displayData"/>
    </html:select>
    When i submit the form, i am able to get the value of the selected Value in the List.
    JUCtrlListBinding listBinding = (JUCtrlListBinding) formBean.get("Lob");
    Row r = (Row) listBinding.getSelectedValue();
    String SoLob = (String) r.getAttribute("LobLob");
    But i am unable to set the selected value back in the list when i return to the form again.
    I tried the following:
    int x = listBinding.getSelectedIndex();
    listBinding.setSelectedIndex(x);
    It would be really helpful if some one could let me know what needs to be done.
    Thanks,
    Subashini

    Gyan,
    I tried with the code you suggest , its trying to set the value of a field, but my requirement is to change the property for
    "Rendered" on the field so that based on search parameters we can hide result table column.
    After searching online, I modified the controller code as follows
    1.Created a new transient attribute (Testrender) with type "Boolean" and always updateable.
    2.changed the "Rendered" property value on the field I want to hide as ${oa.PoSearchVO1.Testrender}.
    3. Modified the controller code as follows :
            OAApplicationModule am = pageContext.getApplicationModule(webBean);
            OAViewObject testvo = (OAViewObject)am.findViewObject("PoSearchVO1");
            OADBTransaction txn = am.getOADBTransaction();
            if (testvo != null)
              OARow row = (OARow)testvo.first();
                if (a.equalsIgnoreCase("Approved"))
                     row.setAttribute("Testrender", Boolean.FALSE);
                else
                     row.setAttribute("Testrender", Boolean.TRUE);
            } But now when i run the page and populate the search field and click "Go" its hiding the field but throwing the following error and the search results are also bad.
    Its using only the value of one search parameter but not the second one .
    The search cannot be executed because the table has pending changes that would be lost. I think the vo is getting dirty when I update the transient attribute value. So I used txn.rollback ,then search is working fine but its not hiding the field.
    So can you please let me know how to proceed from here ?

  • Unable to set  kernel parameters using solaris 10

    HI,
    I'm trying to install DB2 8.2 on solaris 10.the problem i'm encountering in installation is unable to set MSGMAX and MSGMNB kernel parameters. I tried setting the new replacement parameter process.max-msq-qbytes through projects->Resource Control from the management console.but still uinable to set the parameters.
    Can anyone help me in setting the above kernel parameters.
    Its not possible to set it thru /etc/system.
    thanx
    guru

    HI,
    I'm trying to install DB2 8.2 on solaris 10.the problem i'm encountering in installation is unable to set MSGMAX and MSGMNB kernel parameters. I tried setting the new replacement parameter process.max-msq-qbytes through projects->Resource Control from the management console.but still uinable to set the parameters.
    Can anyone help me in setting the above kernel parameters.
    Its not possible to set it thru /etc/system.
    thanx
    guru

  • Unable to set kernel parameters using Reource Control

    HI,
    I'm trying to install DB2 8.2 on solaris 10.the problem i'm encountering in installation is unable to set MSGMAX and MSGMNB kernel parameters. I tried setting the new replacement parameter process.max-msq-qbytes through projects->Resource Control from the management console.but still uinable to set the parameters.
    Can anyone help me in setting the above kernel parameters.
    Its not possible to set it thru /etc/system.
    thanx
    guru

    I finally had some time to find the actual syntax for tuning process.max-msq-qbytes and process.max-msg-messages. prctl will change those paramaters for an existing process.
    As a proof of concept I changed process.max-msg-messages to a 64k. process 2221 is my shell.
    prctl -n process.max-msg-messages  -r -v 64k 2221
    bash-3.00# prctl 2221
    process: 2221: -csh
    NAME    PRIVILEGE       VALUE    FLAG   ACTION                       RECIPIENT
    process.max-port-events
            privileged      65.5K       -   deny                                 -
            system          2.15G     max   deny                                 -
    process.max-msg-messages
            privileged      64.0K       -   deny                                 -
            system          4.29G     max   deny                                 -
    process.max-msg-qbytes
            privileged      64.0KB      -   deny                                 -
            system          4.00GB    max   deny                                 -And now to make this persist across reboots...
    So long as I'm experimenting, I'll use projadd and add a new project which just includes me and not the rest of the users in the group staff.
    bash-3.00# projadd -p 115 -U testuser \
    -K "process.max-msg-messages=(priv,64K,deny)" \
    -K "process.max-msg-qbytes=(priv,64K,deny)" \
    user.me This creates the following /etc/project:
    (note the last line is all on one line in the real file minus the \.)
    placer% cat /etc/project
    system:0::::
    user.root:1::::
    noproject:2::::
    default:3::::
    group.staff:10::::
    user.me:115::testme::\
    process.max-msg-messages=(priv,64000,deny);process.max-msg-qbytes=(priv,65536,deny)After a reboot I checked the values by loging in under my user and running prctl.
    placer% prctl $$
    process: 419: -csh
    NAME    PRIVILEGE       VALUE    FLAG   ACTION                       RECIPIENT
    process.max-port-events
            privileged      65.5K       -   deny                                 -
            system          2.15G     max   deny                                 -
    process.max-msg-messages
            privileged      64.0K       -   deny                                 -
            system          4.29G     max   deny                                 -
    process.max-msg-qbytes
            privileged      64.0KB      -   deny                                 -
            system          4.00GB    max   deny                                 -
    process.max-sem-ops
            privileged        512       -   deny                                 -
            system          2.15G     max   deny                                 -
    ...Solaris 10 has different tunables and default values from 8 and 9. It's default tunables should be good for most applications per a given system's hardware. There are some great blog articles, and discussions on opensolaris.org and blogs.sun.com on tuning that explain the intent of tuning and why we shouldn't have to tune.
    Cheers,
    ~~sa

  • Unable to set kernel parameters using Reource Control in solaris 10

    HI,
    I'm trying to install DB2 8.2 on solaris 10.the problem i'm encountering in installation is unable to set MSGMAX and MSGMNB kernel parameters. I tried setting the new replacement parameter process.max-msq-qbytes through projects->Resource Control from the management console.but still uinable to set the parameters.
    Can anyone help me in setting the above kernel parameters.
    Its not possible to set it thru /etc/system.
    thanx

    HI,
    I'm trying to install DB2 8.2 on solaris 10.the problem i'm encountering in installation is unable to set MSGMAX and MSGMNB kernel parameters. I tried setting the new replacement parameter process.max-msq-qbytes through projects->Resource Control from the management console.but still uinable to set the parameters.
    Can anyone help me in setting the above kernel parameters.
    Its not possible to set it thru /etc/system.
    thanx

  • [CUPS] Unable to set my new CUPS server with an HP printer (solved)

    Hello,
    I come from Gentoo and I just installed my new Arch Linux server on a notebook with 2 NIC ... Now, I'm trying to set the CUPS deamon to share printer (HP DESKJET 895Cxi) for 3 workstations in my LAN. Before, it was working (printer and linux ... has never been very good...). But now, I'm not able, and it's about 1 hour I'm looking for a stupid error that I don't know if it's important or not ...
    (all is on my Wiki : http://pix-mania.dyndns.org/mediawiki/index.php/Cups)
    Here are a list of installed package : cups ghostscript gsfonts hplip
    The printer (USB) :
    (lsusb)
    Bus 002 Device 004: ID 03f0:0004 Hewlett-Packard DeskJet 895c
    (dmesg)
    usb 2-2: new full speed USB device using uhci_hcd and address 4
    usblp0: Disabling reads from problematic bidirectional printer
    usblp0: USB Unidirectional printer dev 4 if 0 alt 1 proto 2 vid 0x03F0 pid 0x0004
    Current configuration (only change made to original configuration) :
    Listen <ip_serveur>:631
    <Location />
    Order allow,deny
    Allow From all
    </Location>
    <Location /admin>
    Order allow,deny
    Allow From all
    </Location>
    <Location /admin/conf>
    AuthType Default
    Require user @SYSTEM
    Order allow,deny
    Allow From all
    </Location>
    When starting cups service, all is fine : I can use a workstation to browse the Web Interface ... but printer seems to be not detected since it tells "No printers." when clicking on "Printers" tab. So, I go to "Administration" tab and, under "Printers", I click the "Find New Printers" and again, it says : "No printers found.".
    Not goot ... so, like I was doing with Gentoo, I click on "Add Printer" ... then, I receive a redirection to HTTPS ... click on the link (or wait) and nothing happend (timed out): web interface is broken for HTTPS ... I really got no idea why HTTPS is not working .
    After that, when updating the configuration (with vim) and restarting the service, deamon won't stop and I have to send a "kill -9" to the cupsd pid ...
    I look in log and I can just see this :
    E [09/Jan/2011:18:19:55 +0100] Unable to set ACLs on root certificate "/var/run/cups/certs/0" - Operation not supported
    I googled for one hour, but is not more working right now ...
    Some questions :
    - the printer is not detected? Is it because of the ACLs errors ?
    - what the hell is "ACLs" for cups ? Never seen that before ...
    So, right now, I got a working HTTP but not working HTTPS, with no printer available ... very useless
    This printer is working "well" with cups, since it's an old model and it's a HP printer, so is "fully?" working on Linux ...
    Any idea on what to do/try now ?
    EDIT: after that message, I just refresh the page (I'm using Chromium) and ... magically, HTTPS was working (after accept untrusted certificate for HTTPS?) ... Then, go back to "Find New Printers" and now, it was working. Printer has been added ... Seems to be OK, but the "Print Self Test Page" doesn't work right now because of this :
    Unsupported format 'application/vnd.cups-command'!
    Current log file :
    E [09/Jan/2011:18:19:55 +0100] Unable to set ACLs on root certificate "/var/run/cups/certs/0" - Operation not supported
    E [09/Jan/2011:18:24:48 +0100] Unable to set ACLs on root certificate "/var/run/cups/certs/0" - Operation not supported
    E [09/Jan/2011:18:29:08 +0100] Unable to set ACLs on root certificate "/var/run/cups/certs/0" - Operation not supported
    E [09/Jan/2011:18:36:28 +0100] Unable to set ACLs on root certificate "/var/run/cups/certs/0" - Operation not supported
    E [09/Jan/2011:18:36:42 +0100] Unable to set ACLs on root certificate "/var/run/cups/certs/0" - Operation not supported
    E [09/Jan/2011:18:41:06 +0100] Unknown SystemGroup "sys root lpadmin" on line 13, ignoring!
    E [09/Jan/2011:18:41:06 +0100] Unable to set ACLs on root certificate "/var/run/cups/certs/0" - Operation not supported
    E [09/Jan/2011:18:42:24 +0100] Unknown SystemGroup "lpadmin" on line 14, ignoring!
    E [09/Jan/2011:18:42:24 +0100] Unable to set ACLs on root certificate "/var/run/cups/certs/0" - Operation not supported
    E [09/Jan/2011:18:45:12 +0100] Unable to set ACLs on root certificate "/var/run/cups/certs/0" - Operation not supported
    E [09/Jan/2011:18:47:32 +0100] Unable to set ACLs on root certificate "/var/run/cups/certs/0" - Operation not supported
    E [09/Jan/2011:18:48:35 +0100] Unable to set ACLs on root certificate "/var/run/cups/certs/0" - Operation not supported
    E [09/Jan/2011:18:49:35 +0100] Unable to set ACLs on root certificate "/var/run/cups/certs/0" - Operation not supported
    E [09/Jan/2011:18:59:46 +0100] Unable to set ACLs on root certificate "/var/run/cups/certs/0" - Operation not supported
    E [09/Jan/2011:19:33:06 +0100] Unable to encrypt connection from <ip_workstation> - A TLS packet with unexpected length was received.
    E [09/Jan/2011:19:33:06 +0100] Unable to encrypt connection from <ip_workstation> - A TLS packet with unexpected length was received.
    E [09/Jan/2011:19:33:06 +0100] Unable to encrypt connection from <ip_workstation> - A TLS packet with unexpected length was received.
    E [09/Jan/2011:19:33:13 +0100] Unable to encrypt connection from <ip_workstation> - A TLS packet with unexpected length was received.
    E [09/Jan/2011:19:33:13 +0100] Unable to encrypt connection from <ip_workstation> - A TLS packet with unexpected length was received.
    E [09/Jan/2011:19:33:13 +0100] Unable to encrypt connection from <ip_workstation> - A TLS packet with unexpected length was received.
    E [09/Jan/2011:19:33:13 +0100] Unable to encrypt connection from <ip_workstation> - A TLS packet with unexpected length was received.
    E [09/Jan/2011:19:33:13 +0100] Unable to encrypt connection from <ip_workstation> - A TLS packet with unexpected length was received.
    E [09/Jan/2011:19:33:37 +0100] [CGI] Unable to create avahi client: No such file or directory
    E [09/Jan/2011:19:33:37 +0100] [cups-deviced] PID 2744 (dnssd) stopped with status 1!
    E [09/Jan/2011:19:33:37 +0100] Unable to encrypt connection from <ip_workstation> - A TLS packet with unexpected length was received.
    E [09/Jan/2011:19:34:09 +0100] [CGI] Unable to create avahi client: No such file or directory
    E [09/Jan/2011:19:34:09 +0100] [cups-deviced] PID 2763 (dnssd) stopped with status 1!
    E [09/Jan/2011:19:39:51 +0100] Unable to encrypt connection from <ip_workstation> - A TLS packet with unexpected length was received.
    E [09/Jan/2011:19:39:52 +0100] [cups-driverd] Bad driver information file "/usr/share/cups/model/foomatic-db-ppds/Kyocera/ReadMe.htm"!
    E [09/Jan/2011:19:41:23 +0100] [cups-driverd] Bad driver information file "/usr/share/cups/model/foomatic-db-ppds/Kyocera/ReadMe.htm"!
    E [09/Jan/2011:19:44:49 +0100] Unable to encrypt connection from <ip_workstation> - A TLS packet with unexpected length was received.
    E [09/Jan/2011:19:47:16 +0100] Returning IPP client-error-document-format-not-supported for Send-Document (ipp://localhost:631/printers/HP_DeskJet_895C_USB_ES89M230MKGO_HPLIP) from localhost
    E [09/Jan/2011:19:49:36 +0100] Returning IPP client-error-document-format-not-supported for Send-Document (ipp://localhost:631/printers/HP_DeskJet_895C_USB_ES89M230MKGO_HPLIP) from localhost
    EDIT2: tried to print "Print Test Page" instead of the above ... and got these errors :
    (web interface)
    "/usr/lib/cups/backend/hp failed"
    (error_log)
    E [09/Jan/2011:20:08:31 +0100] [CGI] Unable to create avahi client: No such file or directory
    E [09/Jan/2011:20:08:31 +0100] [cups-deviced] PID 3126 (dnssd) stopped with status 1!
    D [09/Jan/2011:20:09:16 +0100] [Job 5] perl: warning: Setting locale failed.
    D [09/Jan/2011:20:09:16 +0100] [Job 5] prnt/backend/hp.c 559: dBus Connection Error (Failed to connect to socket /var/run/dbus/system_bus_socket: No such file or directory)!
    D [09/Jan/2011:20:09:16 +0100] [Job 5] perl: warning: Please check that your locale settings:
    D [09/Jan/2011:20:09:16 +0100] [Job 5] LANGUAGE = (unset),
    D [09/Jan/2011:20:09:16 +0100] [Job 5] LC_ALL = (unset),
    D [09/Jan/2011:20:09:16 +0100] [Job 5] LANG = "fr_FR.UTF-8"
    D [09/Jan/2011:20:09:16 +0100] [Job 5] are supported and installed on your system.
    D [09/Jan/2011:20:09:16 +0100] [Job 5] perl: warning: Falling back to the standard locale ("C").
    D [09/Jan/2011:20:09:16 +0100] [Job 5] GhostScript extra search path ('GS_LIB'): /usr/share/cups/fonts
    D [09/Jan/2011:20:09:16 +0100] [Job 5] Pondering option 'job-uuid=urn:uuid:e4395221-1209-32a9-5add-30f7945eaa6b'
    D [09/Jan/2011:20:09:16 +0100] [Job 5] Unknown option job-uuid=urn:uuid:e4395221-1209-32a9-5add-30f7945eaa6b.
    D [09/Jan/2011:20:09:16 +0100] [Job 5] Pondering option 'job-originating-host-name=localhost'
    D [09/Jan/2011:20:09:16 +0100] [Job 5] Unknown option job-originating-host-name=localhost.
    D [09/Jan/2011:20:09:16 +0100] [Job 5] Pondering option 'time-at-creation=1294600145'
    D [09/Jan/2011:20:09:16 +0100] [Job 5] Unknown option time-at-creation=1294600145.
    D [09/Jan/2011:20:09:16 +0100] [Job 5] Pondering option 'time-at-processing=1294600145'
    D [09/Jan/2011:20:09:16 +0100] [Job 5] Unknown option time-at-processing=1294600145.
    D [09/Jan/2011:20:09:16 +0100] [Job 5] Pondering option 'AP_D_InputSlot='
    D [09/Jan/2011:20:09:16 +0100] [Job 5] Unknown option AP_D_InputSlot=.
    D [09/Jan/2011:20:09:16 +0100] [Job 5] prnt/backend/hp.c 625: ERROR: 5021 device communication error!
    D [09/Jan/2011:20:09:16 +0100] [Job 5] GPL Ghostscript 9.00: Unrecoverable error, exit code 1
    D [09/Jan/2011:20:09:16 +0100] [Job 5] Process 3154 ending: "gs -q -dBATCH -dPARANOIDSAFER -dQUIET -dNOPAUSE -sDEVICE=ijs -sIjsServer=hpijs -dDEVICEWIDTHPOINTS=5..."
    D [09/Jan/2011:20:09:16 +0100] [Job 5] renderer return value: 1
    D [09/Jan/2011:20:09:16 +0100] [Job 5] renderer received signal: 1
    D [09/Jan/2011:20:09:16 +0100] [Job 5] Process dying with "Possible error on renderer command line or PostScript error. Check options.", exit stat: 3
    D [09/Jan/2011:20:09:16 +0100] [Job 5] error: No such file or directory (2)
    D [09/Jan/2011:20:09:16 +0100] [Job 5] Cleaning up ...
    D [09/Jan/2011:20:09:16 +0100] [Job 5] Killing process 3153 (KID4) with signal 15
    D [09/Jan/2011:20:09:16 +0100] [Job 5] KID3 exited with status 3
    D [09/Jan/2011:20:09:16 +0100] [Job 5] Renderer exit stat: 3
    D [09/Jan/2011:20:09:16 +0100] [Job 5] Killing process 3153 (KID4) with signal 9
    D [09/Jan/2011:20:09:16 +0100] [Job 5]
    D [09/Jan/2011:20:09:16 +0100] [Job 5] Closing foomatic-rip.
    D [09/Jan/2011:20:09:16 +0100] [Job 5] Renderer process finished
    D [09/Jan/2011:20:09:16 +0100] [Job 5] Process dying with "Error closing renderer", exit stat: 3
    D [09/Jan/2011:20:09:16 +0100] [Job 5] error: (0)
    D [09/Jan/2011:20:09:16 +0100] [Job 5] Cleaning up ...
    D [09/Jan/2011:20:09:16 +0100] [Job 5] Killing process 3152 (KID3) with signal 15
    D [09/Jan/2011:20:09:16 +0100] [Job 5] Killing process 3152 (KID3) with signal 9
    D [09/Jan/2011:20:09:16 +0100] [Job 5] Error closing renderer
    D [09/Jan/2011:20:09:16 +0100] [Job 5]
    D [09/Jan/2011:20:09:16 +0100] [Job 5] Closing foomatic-rip.
    D [09/Jan/2011:20:09:16 +0100] [Job 5] Backend returned status 4 (stop printer)
    D [09/Jan/2011:20:09:16 +0100] [Job 5] Printer stopped due to backend errors; please consult the error_log file for details.
    D [09/Jan/2011:20:09:16 +0100] [Job 5] End of messages
    D [09/Jan/2011:20:09:16 +0100] [Job 5] printer-state=5(stopped)
    D [09/Jan/2011:20:09:16 +0100] [Job 5] printer-state-message="/usr/lib/cups/backend/hp failed"
    D [09/Jan/2011:20:09:16 +0100] [Job 5] printer-state-reasons=paused
    E [09/Jan/2011:20:14:21 +0100] [Job 5] Stopping unresponsive job!
    EDIT3: I tested the ppd from cups (and not hpijs) and got the same problem ... The latest idea was to unplug USB and plug to another USB ... and it seams to works ... Now, the test page print "fine" ...
    So, I shared the printer over my LAN, and tried to print from a workstation ... it works fine now ... when looking error_log, I can see than there is no more Ghostscript error .
    But I have a last question .. if somebody can answer ... When choosing a PPD file, I have the choice between 3 version :
    - hpijs, 3.10.9
    - hpijs, 3.10.9.11
    - hpcups, 3.10.9
    Don't know why onlyl the HPIJS got the ".11" and not any others ... I selected this one and here is what Cups say about the printer :
    HP_DeskJet_895C_USB_ES29Z231JKGO_HPLIP HP DeskJet 895C Local Printer HP Deskjet 895c hpijs, 3.10.9.11 Idle - "ready to print"
    I hope when upgrading the system, I will not get error from PPD file not found ... With my current PPD file, I get (really not sure) this error, but print is working so, it doesn't matter :
    E [09/Jan/2011:21:08:04 +0100] [CGI] Unable to create avahi client: No such file or directory
    E [09/Jan/2011:21:08:04 +0100] [cups-deviced] PID 5705 (dnssd) stopped with status 1!
    E [09/Jan/2011:21:08:09 +0100] [cups-driverd] Bad driver information file "/usr/share/cups/model/foomatic-db-ppds/Kyocera/ReadMe.htm"!
    Thanks for all .. hope you don't think I'm crazy ^^
    Last edited by loopx (2011-01-09 20:31:21)

    WD and their lack of support for Mavericks:
    A recent email from WD reads:
    Dear WD Registered Customer,
    UPDATE: November 1st, 2013
    As a valued WD customer we want to make you aware of new reports of Western Digital and other external HDD products experiencing data loss when updating to Apple's OS X Mavericks (10.9). WD is urgently investigating these reports and the possible connection to the WD Drive Manager, WD Raid Manager and WD SmartWare software applications. Until the issue is understood and the cause identified, WD strongly urges our customers to uninstall these software applications before updating to OS X Mavericks (10.9), or delay upgrading. If you have already upgraded to Mavericks, WD recommends that you remove these applications and restart your computer. If you have already upgraded to Mavericks and are experiencing difficulty in accessing your external hard drive, please do not save anything to the drive, disconnect the drive from your computer, and contact Western Digital Customer Service at http://support.wd.com/country/ for further assistance.
    The WD Drive Manager, WD Raid Manager, and WD SmartWare software applications are not new and have been available from WD for many years. However, solely as a precaution WD has removed these applications from our website as we investigate this issue.
    Sincerely,
    Western Digital
    Which is basically Western Digital informing you that their out-of-date utilities don’t support Mavericks.
    In any case those utilities are not the most effective. The moral of the story is: the first thing you should do after buying a WD drive is to format it as HFS Extended (Journaled) and not install ANY of their utilities but instead use reliable applications like CarbonCopyCloner or SuperDuper - or Time Machine.

  • Unable to set a top-level property object

    Hey guys
    I've made a little sequence to fetch data from a XML file. In that XML file, there is some informations (entries) (IP addresses, port numbers, etc) about 4 different types of devices(sections are the devices names). The first section of the XML tells me the NAMES of the devices I have to configure. The type of each device is located in the entries of each device. Each device has its own information and I can have more than one device of a type. In teststand, I've created custom data types (containers with empty fields) for each type of device I need.
    I want to dynamically create the devices that the XML file tells me to configure as fileglobal variables and populate them with their own config informations (entries of the XML file). I'm creating new fileglobal of the type I want using PropertyObject step and I'm creating an alias of that fileglobal that I'm passing in parameter to a subsequence who's getting infos from XML. Here, everything is OK.
    The problem occurs when I want to configure 2 or more objects of the same type. At the SetProperty step (where i'm creating the fileglobal variable) teststand gives me the error:
    "Unable to set a top-level property object to another object. Error accessing item 'Parameters.DeviceToInitStruct'."
    Is there any ways to avoid that problem and creating to variable of the same type ?
    thanks

    Hi
    I finally got rid of that problem.
    Instead of using a PropertyObject.Clone step, I used a statement who's just copy the "MasterObject" into a "CloneObject" and i'm working with that copy. So the sequence can now create devices of the same type without any conflict with the first device.
    i'll post here a short example to show how i did... maybe it'll be useful for someone...
    regards

  • Microsoft Intune was unable to set the desired mobile device policy for one or more users due to the following error: A2CE0100

    Hi!
    We have fatal or critical error message on Microsoft Intune Portal but all agents are working just fine. Before opening support ticket we would like to hear comments from the experts on this forum. We would also like to fix this error before starting to
    manage mobile devices with Intune.
    Error message on Intune Portal:
    "Microsoft Intune was unable to set the desired mobile device policy for one or more users due to the following error: A2CE0100"
    Repeated: 19 times.
    Class: (System) Policy
    Random Fatal error message on C:\Program Files\Microsoft\OnlineManagement\Logs\PolicyAgent.log found from one Windows 8.1 client:
    2015-02-21 08:49:20:704 2852 1ab0 FATAL: DocumentProvider::IndicateToConsumer/pp->ProcessPolicies(NULL, NULL, NULL, NULL) failed with error 0x800704d5.
    That said, we are not facing any specific problem but we would like to find symptom of this repeating error message on Intune Portal . We would appreciate to get any thoughts about this case.
    Br.
    Jukka

    Hi Jukka,
    Mobile policy doesn't apply to clients using the Full Client download.  Please open a support case so the team can assist in further troubleshooting.
    Thanks,
    Jon L. - MSFT - This posting is provided "AS IS" with no warranties and confers no rights.

  • Unable to open windows using boot camp.  Get message "The bless tool was unable to set the current boot disk."  Any thoughts, Thank you.

    Unable to open windows using boot camp.  Get message "The bless tool was unable to set the current boot disk."   I am using an Imac , Lion operating system, and Windows 7.  It worked a few days ago.  Any thoughts, Thank you.

    Note that nowhere in the Boot Camp instructions does it tell you to use Disk Utility to format the Windows partition. The Boot Camp Assistant program creates the partition & sets the +partition scheme info+ of the disk as appropriate for the Windows installer but the Windows installer itself is responsible for formatting the new partition with the appropriate +file system scheme+ (NTFS for Windows 7).
    If you follow the instructions in the Boot Camp Installation & Setup Guide to the letter you should have no problems installing Windows.

Maybe you are looking for