Smpatch analyze gives different result after disabling Sun Connection Host

Hi all,
after disabling the Sun Connection Hosted functionality due to the EOL announcement I get a smaller number of patches from 'smpatch analyze' for a brand new jumpstarted server (227 patches before vs. 106 patches after).
Trying e.g. a individual analysis for 126651-01 (Synopsis: SunOS 5.10: tar patch) results in
Error: One or more of the updates specified do not have entitlement permissions.
BTW: I ran the analysis for patch 126651-01 before without any problem.
The entitlement file
/var/sadm/spool/cache/entitlement/https%3A%2F%2Fgetupdates1.sun.com%2F%2Fentitlement_client
looks like
----START-ENTITLEMENT-TOKENS----
Solaris10Security
SolarisSecurityUpdates
SolarisDataIntegrityUpdates
SolarisHardwareUpdates
SolarisUtilityUpdates
Public
-----END-ENTITLEMENT-TOKENS-----
And we do have a service plan!
Any ideas?
Kind Regards
Marco

Hi Shen,
thanks for your reply. I tried this procedure from your thread
Try re-registering the patch server after running these commands:
# cacaoadm stop
# cacaoadm status
# rm -r /var/sadm/spool/cache/*
# /usr/lib/cc-ccr/bin/eraseCCRRepository
# rm /var/scn/persistence/SCN*
# cacaoadm start
# /usr/sbin/sconadm register -a -r registrationprofile.properties
But unfortunately the problem persists. The entitlement file looks the same as before. The registrationprofile.properties files looks like
userName=<username>
password=<pw>
hostName=<hostname>
subscriptionKey=<key>
portalEnabled=false
proxyHostName=
proxyPort=
proxyUserName=
proxyPassword=
We use a direct connection to sun's update server without a SunUC proxy. (I disabled the proxy, because I had some trouble with it in the past).
'smpatch get' gives
patchpro.backout.directory - ""
patchpro.baseline.directory - /var/sadm/spool
patchpro.download.directory - /var/sadm/spool
patchpro.install.types - rebootafter:reconfigafter:standard
patchpro.patch.source - https://getupdates1.sun.com/
patchpro.patchset - current
patchpro.proxy.host - ""
patchpro.proxy.passwd **** ****
patchpro.proxy.port - 8080
patchpro.proxy.user - ""

Similar Messages

  • Do IMAQ Cast Image or IMAQ Linear averages give different results when using different computers that are running under Windows XP ?

    Hello
    I'm currently developping an image processing algorithm using Labview 7.1 and the associated IMAQ Vision tools. After several tests, I found a weird result. Indeed, I put the labview algorithm - including the IMAQ VI on the library to get sure that I use all the time the same VI - on my memory stick and used it on two different computers. I tested the same picture (still in my memory stick) and had two very different results.
    After several hours trying to understand why, I found that there were a difference between the results given by both computers at the very begining of the algorithm. Indeed, I used a JPEG file.
    To open it, I first create an Image with IMAQ Create (U8). Then, I open it.
    Then in my first sub-VI, I use IMAQ Cast Image to be sure that the picture is a U8 grayscale picture.
    Right after that, I use the IMAQ Linear Averages. The results of this VI are different on the two computers.
    I tried several time on the same picture : one computer always give me the same result but the two computers give me a different result. So there is no random variable on the results.
    So my question is : Do IMAQ Cast Image or IMAQ Linear averages give different results when using different computers that are running under Windows XP ?
    My bet is on IMAQ Cast Image but I'm not quite sure and I do not undestand why. The labview and IMAQ are the same on both computers.
    The difference between the two computer are above :
    Computer 1 :
    Pentium(R) 4 CPU 3.20GHz with a RAM of 1Go. The processor is an Intel(R).
    The OS is windows XP Pro 2002
    Computer 2 :
    Pentium(R) 4 CPU 2.80GHz with a RAM of 512Mo. The processor is an Intel(R).
    The OS is windows XP Pro 2002.
    If anybody can help me on this problem, it would be really helpful.
    Regards
    Florence P.

    Hi,
    Indeed it's a strange behaviour, could you send me your VI and your JPEG file, (or another file that reproduces) so that I could check this inthere ?
    I'll then try to find out what's happening.
    Regards
    Richard Keromen
    National Instruments France
    #adMrkt{text-align: center;font-size:11px; font-weight: bold;} #adMrkt a {text-decoration: none;} #adMrkt a:hover{font-size: 9px;} #adMrkt a span{display: none;} #adMrkt a:hover span{display: block;}
    >> Découvrez, en vidéo, les innovations technologiques réalisées en éco-conception

  • Filter expression producing different results after upgrade to 11.1.1.7

    Hello,
    We recently did an upgrade and noticed that on a number of reports where we're using the FILTER expression that the numbers are very inflated. Where we are not using the FILTER expression the numbers are as expected. In the example below we ran the 'Bookings' report in 10g and came up with one number and ran the same report in 11g (11.1.1.7.0) after the upgrade and got two different results. The data source is the same database for each envrionment. Also, in running the physical SQL generated by the 10g and 11g version of the report we get different the inflated numbers from the 11g SQL. Any ideas on what might be happening or causing the issue?
    10g report: 2016-Q3......Bookings..........72,017
    11g report: 2016-Q3......Bookings..........239,659
    This is the simple FILTER expression that is being used in the column formula on the report itself for this particular scenario which produces different results in 10g and 11g.
    FILTER("Fact - Opportunities"."Won Opportunity Amount" USING ("Opportunity Attributes"."Business Type" = 'New Business'))
    -------------- Physical SQL created by 10g report -------- results as expected --------------------------------------------
    WITH
    SAWITH0 AS (select sum(case when T33142.OPPORTUNITY_STATUS = 'Won-closed' then T33231.USD_LINE_AMOUNT else 0 end ) as c1,
    T28761.QUARTER_YEAR_NAME as c2,
    T28761.QUARTER_RANK as c3
    from
    XXFI.XXFI_GL_FISCAL_MONTHS_V T28761 /* Dim_Periods */ ,
    XXFI.XXFI_OSM_OPPTY_HEADER_ACCUM T33142 /* Fact_Opportunity_Headers(CloseDate) */ ,
    XXFI.XXFI_OSM_OPPTY_LINE_ACCUM T33231 /* Fact_Opportunity_Lines(CloseDate) */
    where ( T28761.PERIOD_NAME = T33142.CLOSE_PERIOD_NAME and T28761.QUARTER_YEAR_NAME = '2012-Q3' and T33142.LEAD_ID = T33231.LEAD_ID and T33231.LINES_BUSINESS_TYPE = 'New Business' and T33142.OPPORTUNITY_STATUS <> 'Duplicate' )
    group by T28761.QUARTER_YEAR_NAME, T28761.QUARTER_RANK)
    select distinct SAWITH0.c2 as c1,
    'Bookings10g' as c2,
    SAWITH0.c1 as c3,
    SAWITH0.c3 as c5,
    SAWITH0.c1 as c7
    from
    SAWITH0
    order by c1, c5
    -------------- Physical SQL created by the same report as above but in 11g (11.1.1.7.0) -------- results much higher --------------------------------------------
    WITH
    SAWITH0 AS (select sum(case when T33142.OPPORTUNITY_STATUS = 'Won-closed' then T33142.TOTAL_OPPORTUNITY_AMOUNT_USD else 0 end ) as c1,
    T28761.QUARTER_YEAR_NAME as c2,
    T28761.QUARTER_RANK as c3
    from
    XXFI.XXFI_GL_FISCAL_MONTHS_V T28761 /* Dim_Periods */ ,
    XXFI.XXFI_OSM_OPPTY_HEADER_ACCUM T33142 /* Fact_Opportunity_Headers(CloseDate) */ ,
    XXFI.XXFI_OSM_OPPTY_LINE_ACCUM T33231 /* Fact_Opportunity_Lines(CloseDate) */
    where ( T28761.PERIOD_NAME = T33142.CLOSE_PERIOD_NAME and T28761.QUARTER_YEAR_NAME = '2012-Q3' and T33142.LEAD_ID = T33231.LEAD_ID and T33231.LINES_BUSINESS_TYPE = 'New Business' and T33142.OPPORTUNITY_STATUS <> 'Duplicate' )
    group by T28761.QUARTER_YEAR_NAME, T28761.QUARTER_RANK),
    SAWITH1 AS (select distinct 0 as c1,
    D1.c2 as c2,
    'Bookings2' as c3,
    D1.c3 as c4,
    D1.c1 as c5
    from
    SAWITH0 D1),
    SAWITH2 AS (select D1.c1 as c1,
    D1.c2 as c2,
    D1.c3 as c3,
    D1.c4 as c4,
    D1.c5 as c5,
    sum(D1.c5) as c6
    from
    SAWITH1 D1
    group by D1.c1, D1.c2, D1.c3, D1.c4, D1.c5)
    select D1.c1 as c1, D1.c2 as c2, D1.c3 as c3, D1.c4 as c4, D1.c5 as c5, D1.c6 as c6 from ( select D1.c1 as c1,
    D1.c2 as c2,
    D1.c3 as c3,
    D1.c4 as c4,
    D1.c5 as c5,
    sum(D1.c6) over () as c6
    from
    SAWITH2 D1
    order by c1, c4, c3 ) D1 where rownum <= 2000001
    Thank you,
    Mike
    Edited by: Mike Jelen on Jun 7, 2013 2:05 PM

    Thank you for the info. They are definitely different values since ones on the header and the other is on the lines. As the "Won Opportunity" logical column is mapped to multiple LTS it appears the OBI 11 uses a different alogorthim to determine the most efficient table to use in the query generation vs 10g. I'll need to spend some time researching the impact to adding a 'sort' to the LTS. I'm hoping that there's a way to get OBI to use similar logic relative to 10g in how it generated the table priority.
    Thx again,
    Mike

  • P/L A/c for a single day gives different results for different users

    Hi!
         We have SAP B1 8.81 (PL 09) in our office. P/L Account for a particular date gives different value for a single account (Other Income Account) for one user and a consistent value for all other users. The user in question is a super user and does not seem to pick a specific entry posted in the Other Income Account for that day thus giving erroneous result. I have checked the Trial Balance for that day and it gives proper result. Only the P/L account gives erroneous data. Could anyone help with the reason for this.
    Thanks
    Sreekumar

    Hi Jitin,
              Thanks for your reply.
              I tried the way you mentioned. However, the result is the same.
              The entry that I'm talking about is an automatic journal entry posted through the Outgoing Payment option.
              To make it more clear, an outgoing payment to a vendor was posted on 05/19/11 for 1,42,178.32. The check was drawn only for 71,089.16 which was the correct amount and the remaining amount (71,089.16) got posted in the Other Income account. The P/L account for this date is not picking the amount of 71,089.16 in the Other Income Account.
    Thanks & Regards
    Sreekumar
    Edited by: Sreekumar on Jan 31, 2012 5:58 PM

  • Same code gives different results in Matlab Script in Labview and Matlab

    I am implemeting a Matlab code into a LabVIEW application using Matlab script. When I import the exactly same code to the Matlab Script in LabVIEW it gives a different result than it is in Matlab. This code is a simulation code including first kind bessel functions. Using LabVIEW 7.1 and Matlab R14 service pack 3.

    Labview 8.5
    Matlab R2009b
    Attached are the graphs produced by matlab script in labview and in matlab.
    The minimum of graph produced by matlab code is below 1 and that in labview is above 1.
    Thanks a lot for your reply.
    Sorry, I haven't quantified the "sometimes" yet. 
    Attachments:
    matlabsResult.jpg ‏29 KB
    LV.png ‏84 KB

  • Why not exists and minus give different result

    HI,
    I am writing thes query by using two different operators they should give me the same output but they dont. Please can anybody explain why they give different output.
    query 1
    SQL> SELECT COUNT(*) FROM
    2 (select orig_idx
    3 from rel15_au_poi a
    4 MINUS
    5 SELECT ORIG_IDX
    6 FROM REL14_1_AU_POI_INTERM B) ;
    COUNT(*)
    244312
    query 2
    SQL> select count(*)
    2 from rel15_au_poi a
    3 where not exists
    4 (select null from rel14_1_au_poi_interm b where a.orig_idx=b.ORIG_IDX) ;
    COUNT(*)
    245341
    best Regards,

    Because you have duplicates in rel15_au_poi which are counted with exists and removed with minus.
    http://download.oracle.com/docs/cd/B19306_01/server.102/b14200/queries004.htm#sthref3147
    MINUS Example
    The following statement combines results with the MINUS operator, which returns only unique rows returned by the first query but not by the second:
    SELECT product_id FROM inventories
    MINUS
    SELECT product_id FROM order_items;

  • Disk Utility Gives Differing Results...........

    I used disk utility (Leopard, not from start up disk) to repair permissions and to verify the hard drive. Permissions repair worked fine (but v slowly), but verifying the disk brought up an error saying that the disk needed to be repaired. I therefore ran disk utility from the disk (ie holding c at startup etc....).
    Running disk utility eventually resulted in green text saying that no repair of the disk was necessary.
    I restarted, re-ran disk utility in Leopard and it still says the disk needs to be repaired. Onyx also says the disk needs to be repaired. I have run the extended hardware test, which said there are no problems.
    Can anyone shed any light on why I am getting these different results and if I need to be worried???
    Many thanks in advance!

    There are a few important points to verify to ensure an external drive is bootable. Attach the external drive while started up normally. Run Disk Utility. Select the external drive in the left sidebar and look at the info at the bottom of the window. To be able to install Leopard on it, the Partition Map Scheme must be GUID Partition Table.
    If it is GUID Partition Table, you can start up from the Mac OS X installation disk and select it as the target for installation. If the drive is not empty, the data already on it will not be affected, but you should probably back it up if it is valuable. After the installation, start up from the external drive and run Software Update to get all the updates. You will have to run Software Update a few times to get all the inter-dependent updates.
    If it is not GUID Partition Table, you will need to re-partition it and change it to GUID Partition Table using Disk Utility's Partition tab. In this case, the drive will be erased.
    Once you have done the installation and all the updates, you can run Disk Utility while started up from the external drive and select the internal drive for the Repair routine. Since the Repair process should be identical, hopefully the results will be consistent.

  • Same C++ code gives different result in MacOSX and Ubuntu

    Hi
    I have a C++ program which starts by reading an external file, but changing the name of the file give me different results.
    I selected just the important part and made a simple file with the reading part of my program and can be downloaded here
    http://dl.dropbox.com/u/664351/files.zip
    It just reads the external file and does the output of some values.
    It has two external file examples. If one uses "x.txt" everything works fine, but if we use "dados.txt", which is a copy of "x.txt", the last value in the output is wrong.
    The stranger is that if i use these same files in ubuntu everything works just fine!
    I tried in a MacBook, a MacBook Pro and an iMac (all of them with macosx 10.6) and i always got the error, but not on ubuntu!
    Does anyone has a clue about what is happening here?
    Cheers,
    Marcelo

    marcelobarbosa wrote:
    Does anyone has a clue about what is happening here?
    Change OutFile from a 20 byte character array to a string. GCC isn't able to tell what type that is and is picking int or something, corrupting your memory.

  • Matlab 'place' command gives different result in Mathscript/Labview

    Hello,
    I am trying to implement a order reduced observer in Labview. First I used Labview blocks for programming but it happens that I get different results when I am using Matlab. For that reason I used Mathscript to make a simple 'place' command to check it. The result in Labview and Mathscript is the same, but not in Matlab...
    This is the instruction in Matlab:
    A22=[0.9024 0 0;0 0.8607 0; 0 0 0];
    A12=[0 0 0 ;0 0 0.0548;0.0417 0 0;0 -0.0417 0];
    K=place(A22',A12',[0.32 0.31 0.3]);
    Ko=K'
    Ko =
             0         0   13.9664         0
             0         0         0  -13.2062
             0   -5.4745         0         0
    And I have attached the VI where I do the test with Labview blocks and Mathscript. As you can see the result is not even close to be similar. Why is this produced? How can I fix it?
    Thank you a lot in advance!
    Attachments:
    pruebararamathscript.vi ‏39 KB

    Hi,
    Indeed there are a lot of gains that places the poles on the desired locations. BUT, seeing as the Varga's SEAP algorithm is implemented, the G matrix is selected through a uniform white noise generator and the desired characteristics are verified (controllability). This does not unsure optimization regarding stability margins or disturbance rejection (sensitivity minimization regarding perturbations an matrix A or B).
    As an example, the system presented in this thread, the algorithm implemented in CD Pole Placement.vi computes an observer gain that determine a 7,4455 value for closed loop infinity norm. The place algorithm in Matlab computes an observer gain that determine a 2,8230 value for closed loop infinity norm, so better noise rejection or robustness.
    cosmin

  • Why jpeg file exported from Aperture with 300 dpi's it opens with 72 dpi's on PS ( image size)? I've tried several combinations and all give different results. And I am confused on what is the best workflow for me.

    Ok.
    My workflow is
    RAW > Aperture Library > export jpeg high resolution 300 dpi's > one file on PS > edit > Save us jpeg. The I realised that files from PS were being save in a smaller size from the ones exported form Aperture. That is when I went o see <image size> on PS and files were with 72.
    Since ten I've trying different things
    1. Aperture > export as PSD > open on PS > edit > Save us jpeg = small file (around 15 MB)
    2. Aperture > export as jpeg high resolution 300 dpi's > open on PS > CHANGE dpi's to 300 on <image size> edit > Save us jpeg = big file (Huge, actually)
    What am I doing wrong? Would someone give me guidance and tell me what is the best workflow, considering I edit photos to deliver to my clients and I shall give them 300 dpi's.
    Thank you

    This is a known Aperture issue: Problem with Aperture 3.6 preset exports. | Apple Support Communities
    Benjamin

  • Calculation in owb gives different result vs db

    hello
    i have a case when calculation in expression in a mapping
    when i run the mapping the result of this expression is different from when i run it directly at the database level
    has anyone ever encountered a situation like this?
    i appreciate any tips
    rgds

    hello
    i found the cause of that;
    apparently the date format set at db level was dd-rrrr-yy, whereas when running the same expression via sqldeveloper i had the local setting for date :dd-mm-rrrr
    as such when the calculation was executied via OWB it used the defualt nls setting and resulted in incorrect outcome;
    rgds

  • Insert, select and create table as give different results

    Hi all
    I have a strange situation with this three cases:
    1) select statement: SELECT (...)
    2) insert statement with that same select as above: INSERT INTO SELECT (...)
    3) create table statement again with that same select: CREATE TABLE AS SELECT (...)
    Each of these cases produce different number of rows (first one 24, second 108 and third 58). What's more the data for second and third case doesn't have any sense with what they should return. The first case returns good results.
    One interesting thing is that select uses "UNION ALL" between 2 queries. When simple UNION is used, everything works fine and all three cases return 24 rows. Also if each query is run seaprately, they work fine.
    Anyone encountered something like this? (before i create an SR :)
    Database is 10.2.0.2 on AIX 5.3. It's a data warehouse.
    Edited by: dsmoljanovic on Dec 10, 2008 3:57 PM

    I understand UNION vs UNION ALL. But that doesn't change the fact that same SELECT should return same set of rows wether in INSERT, CREATE TABLE AS or a simple SELECT.
    DB version is 10.2.0.2.
    Here is the SQL:
    INSERT INTO TMP_TRADING_PROM_BM_OSTALO
    select
    5 AS VRSTA_PLANIRANJA_KLJUC, u1.UNOS_KLJUC, t1.TRADING_TIP_KLJUC, i1.IZVOR_KLJUC, m1.ITEMNAME AS MJESEC,
    l1.PLAN AS IZNOS, l1.TEKUA AS TEKUCA, l1.PROLA AS PROSLA, l1.PLANTEKUA_ AS IZNOS_TEKUCA, l1.PLANPROLA_ AS IZNOS_PROSLA, l1.TEKUAPROLA_ AS TEKUCA_PROSLA, l1.DATUM_UCITAVANJA
    from
    HR_SP_PLAN.L_12_ET_PROMETI_I_BRUTO_MARZA l1,
    select
    m1.ITEMIID, m1.ITEMNAME
    from
    HR_SP_PLAN.L_12_IT_4_MJESECI m1
    where
    UPPER (m1.ITEMNAME) NOT LIKE '%KVARTAL%' AND UPPER (m1.ITEMNAME) NOT LIKE '%GODINA%' AND UPPER (m1.ITEMNAME) NOT LIKE '%PROC%' and
    m1.DATUM_UCITAVANJA = to_date('24.11.2008','dd.mm.yyyy')
    union all
    select -99, null from dual
    ) m1,
    HR_SP_PLAN.L_12_IT_5_SEKTORI l2,
    HR_SP_PLAN.L_12_IT_2_TIPOVI_OSTALO l3, HR_SP_PLAN.D_UNOS u1, HR_SP_PLAN.D_TRADING_TIP t1, HR_SP_PLAN.L_12_IT_1_PROMET_I_BM_OSTALO p1, HR_SP_PLAN.D_IZVOR i1
    where
    l1.ELIST = l2.ITEMIID and
    l2.ITEMNAME = u1.UNOS_NAZIV and u1.USER_KLJUC = 12 and l2.DATUM_UCITAVANJA = to_date('24.11.2008','dd.mm.yyyy') and
    l1.DIMENSION_1_PROMET = p1.ITEMIID and
    p1.ITEMNAME = i1.IZVOR_NAZIV and i1.USER_KLJUC = 12 and p1.DATUM_UCITAVANJA = to_date('24.11.2008','dd.mm.yyyy') and
    nvl(l1.DIMENSION_4_MJESEC , -99) = m1.ITEMIID and
    l1.DIMENSION_2_TIPOVI = l3.ITEMIID and
    l3.ITEMNAME = t1.TRADING_TIP_NAZIV and l3.DATUM_UCITAVANJA = to_date('24.11.2008','dd.mm.yyyy') and
    l1.DATUM_UCITAVANJA = to_date('24.11.2008','dd.mm.yyyy') and
    'PROC' = 'PLAN'
    union all
    select
    4 AS VRSTA_PLANIRANJA_KLJUC, u1.UNOS_KLJUC, t1.TRADING_TIP_KLJUC, i1.IZVOR_KLJUC, m1.ITEMNAME AS MJESEC,
    l1.PROCJENA AS IZNOS, l1.TEKUA AS TEKUCA, l1.PROLA AS PROSLA, l1.PROCJENATEKUA_ AS IZNOS_TEKUCA, l1.PROCJENAPROLA_ AS IZNOS_PROSLA, l1.TEKUAPROLA_ AS TEKUCA_PROSLA, l1.DATUM_UCITAVANJA
    from
    HR_SP_PLAN.L_13_ET_PROMETI_I_BRUTO_MARZA l1,
    select
    m1.ITEMIID, m1.ITEMNAME
    from
    HR_SP_PLAN.L_13_IT_4_MJESECI m1
    where
    UPPER (m1.ITEMNAME) NOT LIKE '%KVARTAL%' AND UPPER (m1.ITEMNAME) NOT LIKE '%GODINA%' AND UPPER (m1.ITEMNAME) NOT LIKE '%PROC%' and
    nvl(ceil(to_number(m1.ITEMNAME)/3), mod(4, 5)) = mod(4, 5) and m1.DATUM_UCITAVANJA = to_date('24.11.2008','dd.mm.yyyy')
    union all
    select -99, null from dual
    ) m1,
    HR_SP_PLAN.L_13_IT_5_SEKTORI l2, HR_SP_PLAN.L_13_IT_2_TIPOVI_OSTALO l3,
    HR_SP_PLAN.D_UNOS u1, HR_SP_PLAN.D_TRADING_TIP t1,
    HR_SP_PLAN.L_13_IT_1_PROMET_I_BM_OSTALO p1, HR_SP_PLAN.D_IZVOR i1
    where
    l1.ELIST = l2.ITEMIID and
    l2.ITEMNAME = u1.UNOS_NAZIV and u1.USER_KLJUC = 13 and l2.DATUM_UCITAVANJA = to_date('24.11.2008','dd.mm.yyyy') and
    l1.DIMENSION_1_PROMET = p1.ITEMIID and
    p1.ITEMNAME = i1.IZVOR_NAZIV and i1.USER_KLJUC = 13 and p1.DATUM_UCITAVANJA = to_date('24.11.2008','dd.mm.yyyy') and
    nvl(l1.DIMENSION_4_MJESEC , -99) = m1.ITEMIID and
    l1.DIMENSION_2_TIPOVI = l3.ITEMIID and
    l3.ITEMNAME = t1.TRADING_TIP_NAZIV and l3.DATUM_UCITAVANJA = to_date('24.11.2008','dd.mm.yyyy') and
    l1.DATUM_UCITAVANJA = to_date('24.11.2008','dd.mm.yyyy') and
    'PROC' = 'PROC';

  • Custom metadata properties give no results after upgrade from 2010

    Hi,
    Although I see my custom managed properties in my Search Schema in Central Admin, I am unable to find results using them e.g. 'VendorName:Marcel'. I have tried changing the column value in a document and performed a full crawl, but still no results. General
    searches such as 'Marcel' work fine and even something like 'author:marcel' (which is a built in managed property) works too. What am I missing? How can I determine the issue? Nothing stands out in the ULS logs.
    Background:
    I recently upgraded my managed metadata service and search service from 2010 to 2013.
    I then upgraded the content db holding my root site collection, and had to re-create my search service proxy to get searches to work. I then upgraded the content db which holds my content hub site collection. This includes site columns, many of which
    are also Managed Properties.
    Crawls are running with 8 errors - none of which are for the document I changed. Warnings are mainly about 'png' files.
    In 2010 I had scopes, which I no longer see in Central Admin. Although I have converted the content database without error, I have not converted the Site Collection to 2013 yet.
    macrel

    I had to set each metadata property 'queryable' property and then run a full crawl after which the results were returned. I am surprised that this did not come over from the upgrade of the search service.
    macrel

  • Query via linked server to DB2/NT gives different results

    I have a DB2 9.7 database I need to query. I'm testing from four separate SQL Servers.
    ServerA - SQL 2012 SP2
    ServerB - SQL 2012 SP1
    ServerC - SQL 2012 SP1 CU1
    ServerD - SQL 2012 SP1
    I set up a linked server from ServerA. I run QueryA, and am happy with the results.
    I create a linked server from ServerB. I run QueryA and get the message, "Error converting data type DBTYPE_DBDATE to datetime."
    I did the same test on ServerC. It works fine.
    If, inside OPENQUERY, I convert the datetime to a varchar, call it QueryB, it mostly works, but will be missing a small percentage of datetime values from that converted column when run on ServerB. If I run QueryB on the on ServerA, those missing
    values are there. Same for ServerC.
    At this point, I can think only that the main difference is the service pack and CU level. I can't find any documentation on updates that are relevant, but doesn't mean much.
    Then, I did the same test on ServerD. ServerD is at the same SP/CU level as ServerB. ServerD worked just fine. Apparently it's not the service pack or cumulative update level making the difference.
    My setup: On each server I loaded MS OLE DB Provider for DB2 Version 4.0. I used the same sp_addlinkedserver script with the same connection string on each server. QueryA and QueryB both used OPENQUERY. They were identical except for the CAST to varchar
    for a datetime field. This CAST became a call to a function VARCHAR on the DB2 server in later tests with the same result. 
    Of course, the only place I actually need this working is ServerB.
    tia,
    Steve

    Sure, with some obfuscation and shortening.
    On each server, (except the first where I had to figure it out), I went through the same process.
    Load MS OLE DB Provider for DB2 Version 4.0 version 9.0.1390.0
    Run the Create Linked Server query
    Run QueryA.
    Only on ServerB has this been a problem. 
    --Create Linked Server
    EXEC sp_addlinkedserver
    @server='LinkedServer',
    @srvproduct='Microsoft OLE DB Provider for DB2',
    @catalog='DBName',
    @provider='DB2OLEDB',
    @provstr='Provider=DB2OLEDB;User ID=Username;Password=Password;Initial Catalog=DBName;Network Transport Library=TCPIP;Host CCSID=1252;PC Code Page=1252;Network Address=127.0.0.1;Network Port=50000;Package Collection=MSDB2COL;Process Binary as Character=False;Units of Work=RUW;DBMS Platform=DB2/NT;Use Early Metadata=False;Defer Prepare=False;DateTime As Char=False;Rowset Cache Size=0;Datetime As Date=False;AutoCommit=True;Authentication=Server;Decimal As Numeric=False;Derive Parameters=True;LoadBalancing=False;Persist Security Info=True;Cache Authentication=False;Mode=Read;Connection Pooling=False;'
    --QueryA
    SELECT * FROM OPENQUERY(LinkedServer,'
    SELECT [ARRIVAL]
          ,[ACCT_CODE]
          ,[ACC_EVENT_ID]
          ,[TRANS_DATET]
          ,[AVAIL_DATE]
          ,[AVAIL_STATUS]
      FROM [DBName].[UserName].[TableName]
    ') AS TableNameLink
    --QueryB
    SELECT * FROM OPENQUERY(LinkedServer,'
    SELECT [ARRIVAL]
    ,[ACCT_CODE]
    ,[ACC_EVENT_ID]
    , CAST([TRANS_DATET] AS VARCHAR(26)) TRANS_DATET
    ,[AVAIL_DATE]
    ,[AVAIL_STATUS]
    FROM [DBName].[UserName].[TableName]
    ') AS TableNameLink
    --QueryC
    SELECT * FROM OPENQUERY(LinkedServer,'
    SELECT [ARRIVAL]
    ,[ACCT_CODE]
    ,[ACC_EVENT_ID]
    , VARCHAR([TRANS_DATET]) TRANS_DATET
    ,[AVAIL_DATE]
    ,[AVAIL_STATUS]
    FROM [DBName].[UserName].[TableName]
    ') AS TableNameLink

  • XML DB: using default namespace gives different result

    In Oracle9i R2 I created an XMLTYPE table based on a XML schema. The schema uses a namespace. When I insert an instance document that has a namespace prefix for every element querying works fine. When I insert an instance document that has the namespace defined as the default namespace, the query returns an empty row for that document. It seems to me that before storing the document the namespaces must be expanded and hence there should be no difference between the two documents and the query should return a value for both rows. Any thoughts? Thanks in advance! Below is the script and the results of the queries.
    ==========
    THE SCRIPT
    ==========
    DROP TABLE islands_xsd
    BEGIN
    DBMS_XMLSCHEMA.DELETESCHEMA('http://www.cumquat.nl/xsd/island.xsd', DBMS_XMLSCHEMA.DELETE_CASCADE);
    END;
    BEGIN
    DBMS_XMLSCHEMA.REGISTERSCHEMA('http://www.cumquat.nl/xsd/island.xsd',
    '<?xml version="1.0"?>
    <schema targetNamespace="http://www.cumquat.nl/xsd/island.xsd"
    xmlns:isl="http://www.cumquat.nl/xsd/island.xsd"
    xmlns="http://www.w3.org/2001/XMLSchema">
    <element name="island" type="isl:islandType"/>
    <complexType name="islandType">
    <sequence>
    <element name="name" type="string"/>
    <element name="country" type="string"/>
    <element name="total_area" type="positiveInteger"/>
    <element name="shoreline" type="positiveInteger"/>
    <element name="pop_density" type="integer"/>
    <element name="dist_continent" type="integer"/>
    </sequence>
    </complexType>
    </schema>');
    END;
    CREATE TABLE islands_xsd OF XMLTYPE
    XMLSCHEMA "http://www.cumquat.nl/xsd/island.xsd"
    ELEMENT "island"
    INSERT INTO islands_xsd VALUES (XMLTYPE(
    '<?xml version="1.0"?>
    <isl:island xmlns:isl="http://www.cumquat.nl/xsd/island.xsd"
    xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance"
    xsi:schemaLocation="http://www.cumquat.nl/xsd/island.xsd http://www.cumquat.nl/xsd/island.xsd">
    <isl:name>Aldabra</isl:name>
    <isl:country>Seychelles</isl:country>
    <isl:total_area>13</isl:total_area>
    <isl:shoreline>104</isl:shoreline>
    <isl:pop_density>0</isl:pop_density>
    <isl:dist_continent>6</isl:dist_continent>
    </isl:island>'))
    INSERT INTO islands_xsd VALUES (XMLTYPE(
    '<?xml version="1.0"?>
    <island xmlns="http://www.cumquat.nl/xsd/island.xsd"
    xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance"
    xsi:schemaLocation="http://www.cumquat.nl/xsd/island.xsd http://www.cumquat.nl/xsd/island.xsd">
    <name>Amsterdam</name>
    <country>France</country>
    <total_area>9</total_area>
    <shoreline>51</shoreline>
    <pop_density>0</pop_density>
    <dist_continent>30</dist_continent>
    </island>'))
    SELECT i.EXTRACT('/isl:island/isl:name', 'xmlns:isl="http://www.cumquat.nl/xsd/island.xsd"')
    FROM islands_xsd i
    SELECT i.EXTRACT('/isl:island/isl:name', 'xmlns:isl="http://www.cumquat.nl/xsd/island.xsd"').GETSTRINGVAL()
    FROM islands_xsd i
    SELECT i.EXTRACT('/isl:island/isl:name/text()', 'xmlns:isl="http://www.cumquat.nl/xsd/island.xsd"').GETSTRINGVAL()
    FROM islands_xsd i
    =======================================
    THE RESULTS OF RUNNING THE QUERIES
    =======================================
    SQL> SELECT i.EXTRACT('/isl:island/isl:name', 'xmlns:isl="http://www.cumquat.nl/xsd/island.xsd"')
    2 FROM islands_xsd i
    3 /
    I.EXTRACT('/ISL:ISLAND/ISL:NAME','XMLNS:ISL="HTTP://WWW.CUMQUAT.NL/XSD/ISLAND.XS
    XMLTYPE()
    2 rows selected.
    SQL> SELECT i.EXTRACT('/isl:island/isl:name', 'xmlns:isl="http://www.cumquat.nl/xsd/island.xsd"').GE
    TSTRINGVAL()
    2 FROM islands_xsd i
    3 /
    I.EXTRACT('/ISL:ISLAND/ISL:NAME','XMLNS:ISL="HTTP://WWW.CUMQUAT.NL/XSD/ISLAND.XS
    <isl:name xmlns:isl="http://www.cumquat.nl/xsd/island.xsd">Aldabra</isl:name>
    2 rows selected.
    SQL> SELECT i.EXTRACT('/isl:island/isl:name/text()', 'xmlns:isl="http://www.cumquat.nl/xsd/island.xs
    d"').GETSTRINGVAL()
    2 FROM islands_xsd i
    3 /
    I.EXTRACT('/ISL:ISLAND/ISL:NAME/TEXT()','XMLNS:ISL="HTTP://WWW.CUMQUAT.NL/XSD/IS
    Aldabra
    2 rows selected.

    This seems wrong to me. I've filed Bug 2400119 to get this looked into by development.
    I'd expect any document with an appropriate qualified name to be found by your XPath expression, irrespective of whether syntactically one of the documents happened to use the default namespace syntax.

Maybe you are looking for

  • Create fully automated Outlook profile at user first time logon to the windows

    Hi experts, I do have one scenario and its quite urgent... I have one email address, that to be act as a notification email from Outlook. The thing is we need to have Outlook fully setup with profile and password when an user log in. I have done the

  • Adding a user in Active Directory

    Hi fellows, I am having a serious problem in creating a new user in active directory. i am using LDAP JNDI code. I can delete and update users attributes, but fail to create users. ctx.createSubcontext("newuser,full domain", attributes); when i speci

  • Quality of .nef LR files Degrading after Edited in CS4 and saved as .tif?

    After I edit a .nef picture in LR, I right click  and select "edit in CS4".  I do my edits in CS4 (usually cloning) and I choose to save the image in CS4, it automatically appears as a .tif in Lightroom with the same rating/color coding that I gave t

  • Safari Problems with Apple Discussions 

    Im Using Camino now, just becuase whenever i log on to Apple disscussions, safari redirects me to the account renewal in apple discussions. When in safari it doesnt remebers my question or subscribtions, which is totally different than camino. Any He

  • Think my drive is failing............

    I have two internal drives but I think my system drive is failing as my ppc g5 tower is going to the grey restart screen at times. So I want to put a new drive in maybe 1tb. I've never done this before but I guess I will take out my second drive, put