Column total_retail bad value

Hi All,
I have a problem with the values on columns TOTAL_RETAIL and UNITS from table IF_TRAN_DATA.
When I consult using PL/SQL or SqlDeveloper, this is the results:
select total_retail, units from rms12.if_tran_data where rownum <= 5;
-0,0101     10
-0,0101     10
-77,34     -1
-66,6724     -1
-77,34     -1
But, when I consult using SQLPlus, this is the results:
select total_retail, units from analize_if_tran_data;
TOTAL_RETAIL      UNITS
*.0000 10*
*.0000 10*
-77.34         -1
-66.6724         -1
-77.34         -1
*5 rows selected.*
The big problem occurs is when I do a calculation with these fields:
select total_retail, units, total_retail/units from analize_if_tran_data;
TOTAL_RETAIL      UNITS TOTAL_RETAIL/UNITS
*.0000 10 .d067195*
*.0000 10 .d067195*
-77.34         -1              77.34
-66.6724         -1            66.6724
-77.34         -1              77.34
*5 rows selected.*
Does anyone know what the problem?
Did anyone see it any time?
Thanks in advance,
Braga

Hi Erik,
Sorry about that. I forgot to explain me well.
Bellow the explain better.
-- Describe of table
SQL> desc analize_if_tran_data
Name Null? Type
TOTAL_RETAIL NUMBER(20,4)
UNITS NOT NULL NUMBER(12,4)
-- Results with bad values
SQL>
SQL> select total_retail, units, total_retail/units from analize_if_tran_data;
TOTAL_RETAIL UNITS TOTAL_RETAIL/UNITS
*.0000 10 .d067195* +(crazy result)+
*.0000 10 .d067195* +(crazy result)+
-77.34 -1 77.34
-66.6724 -1 66.6724
-77.34 -1 77.34
5 rows selected.
-- Inserting a new row with correct value
SQL> insert into analize_if_tran_data values (-0.0101,10);
1 row created.
SQL> commit;
Commit complete.
-- Results with bad and one right value
SQL> select total_retail, units, total_retail/units from analize_if_tran_data;
TOTAL_RETAIL UNITS TOTAL_RETAIL/UNITS
*.0000 10 .d067195* +(crazy result)+
*.0000 10 .d067195* +(crazy result)+
-77.34 -1 77.34
-66.6724 -1 66.6724
-77.34 -1 77.34
-.0101         10            -.00101 +(right result)+
6 rows selected.
Now, see the results on a graphical tool:
(Toad/SqlDeveloper)
select total_retail, units, total_retail/units from analize_if_tran_data;
-0,0101     10     1,9767195 +(crazy result)+
-0,0101     10     1,9767195 +(crazy result)+
-77,34     -1     77,34
-66,6724     -1     66,6724
-77,34     -1     77,34
-0,0101     10     -0,00101 +(correct result)+
I don't know if stayed clear for you.
The right value is '-0,0101', but I only see this value when I execute a query on a graphical tool (pl/sql, sqldeveloper or toad).
If I execute the query the division TOTAL_RETAIL/UNIS is wrong.
I think that now I explained better.
If you see it any time, please help me.
Thank you very much,
Braga

Similar Messages

  • Matrix Bad Value exception in ItemCode column

    hello ,
    i faced an exception with filling itemcode column
    As u know this column is linked object and have CFL (choose from list)
    the problem i faced that on choosing from list i choose an item to add new line in matrix
    this event done and a new line added sucessfully
    but the problem was with CFL , that it didnt able to close , it still opened and you need to press ESC or click on cancel to close it
    when i tracking the exception i found "Bad value" exception, although the itemcode column filled but the
    also i bind all matrix columns with data sources and i can fill matix at first , but when i want to add new line i faced this problem
    best regards and thank you in advance ....
    Ammar

    Thanks Rasmus,
    But that's something I've already tried - it doesn't seem to make any difference.

  • Form - Bad value (Goods Issue Form)

    Hello,
    I'm trying to fill Goods Issue form programatically. But when I try to fill some values into the matrix I got two errors-
    Referenced work order does not exist.
    Form - Bad Value
    I can enter the same values by hand, it works without any problem.
    One more thing. Let's say i got master data items with indexes 123, 1024. When I try to enter these values programmatically into Good Issue matrix (into index column) I got above messages. But when I the same with just value 1 it fills row with index 123.
    What is happening?
    Thanks in advance
    Wydrzycki Kamil

    hello freind
    it may be due to refferd linked key mens primary key.pls cheak when yu insert data in database first cheak this entry may be linked with any primary key

  • Exception Form Bad Value

    Hi Experts,
    I am trying to assign the value of warehouse column of a sap document by code. Generally we have cfl to choose warehouse.
    So the problem is that it is taking the value by code but throwing the exception .The message is "Form Bad Value".
    The Code is like
                                                            EditText oItemEdit = (EditText)oMatrix.Columns.Item("10").Cells.Item(i).Specific;
                                                            string whse = "07";
                                                            oItemEdit.Value = whse.Trim();
    Thanks
    Amit

    Hi Amit,
    Try This....
    If pVal.FormType = "940" And pVal.ColUID = "1" And pVal.EventType = SAPbouiCOM.BoEventTypes.et_LOST_FOCUS And pVal.BeforeAction = False Then
                Try
                    oform = sbo_application.Forms.GetFormByTypeAndCount(pVal.FormType, "1")
                    Dim omat As SAPbouiCOM.Matrix
                    omat = oform.Items.Item("23").Specific
                    Dim whs As String
                    whs = "07"
                    omat.Columns.Item("5").Cells.Item(pval.row).Specific.value = whs.Trim
                Catch ex As Exception
                    sbo_application.MessageBox(ex.Message)
                End Try
            End If
    Its working fine in my system
    Thanks
    Shafi
    Edited by: shafi_sunshine on Sep 20, 2011 9:39 AM

  • Error BEA-382510 bad value for type node

    Hi,
    We have a requirement to read the xpath from the resource file placed in project folder in OSB and use that xpath to generate the report. For this I have done the following:
    1) Made the xpath entry to properties file as <xpath>$body/*/*/EMPTYPE/EMPDATA/EMPNO</xpath>
    2) Created a variable to fetch the xpath value from property file, say xpathvar
    3) Created another variable, say reportvar which will have the contents: <report>{$xpathvar/text()}</report>. This should fetch the xpath from the variable and replace it in the calling part to fetch the value of the xpath.
    4) For report generation the in key value it is ./text() from variable reportvar.
    When I test this it throws error:
    <con:errorCode>BEA-382510</con:errorCode>
    <con:reason>
    OSB Assign action failed updating variable "reportxpath": com.bea.wli.common.xquery.XQueryException: Error parsing XML: line 1, column 19: {err}XP0006: "$body/*/*/EMPTYPE/EMPDATA/EMPNO ({http://www.w3.org/2001/XMLSchema}string)": bad value for type node
    </con:reason>
    Any pointers will be highly appreciated.
    Thanks!

    I think you cant substitute xpath as you have did.
    As shown in the blog which you mentioned in the other post you may need to have a xslt transformation with xalan:evaluate to correctly process the xpath read from the properties file.
    <report>{$xpathvar/text()}</report>. Here $xpathvar having the value '$body/*/*/EMPTYPE/EMPDATA/EMPNO' will be treated as string type (this will be string as you have read from the properties file as string) and not node. And you are trying to apply the text() function on a string instead of node which is resulting in that error.

  • Write to Measurement File time column wrong starting value

    I am using Write tp Measurement File express VI.  I select to write one time column.  The values writtent in the time column start from a value that does not seem to correspond to anything in the program -- sometimes as much as 2-3 days before the saving begins.
    Can I force the starting value  of time to something, or get it to write the time since we started saving, or something predictable?

    Include code whenever available.  Asking somebody to troubleshoot your software without any code is like asking your mechanic to fix your car without taking your car to him.
    Source
    I don't recommend using the Write Express VI.  I recommend using the File I/O primatives because they allow for much more control over what gets written and how.  Still I would expect the Express VI to be able to write time properly so I'd assume there is some kind of bug with the data being given to the function.  Post some code so we can see what is going on.
    Unofficial Forum Rules and Guidelines - Hooovahh - LabVIEW Overlord
    If 10 out of 10 experts in any field say something is bad, you should probably take their opinion seriously.

  • Mkfs: bad value for nbpi: must be at least 1048576 for multi-terabyte, nbpi

    Hi, guys!
    *1. I have a big FS (8 TB) on UFS which contains a lot of small files ~ 64B-1MB.*
    -bash-3.00# df -h /mnt
    Filesystem Size Used Avail Use% Mounted on
    /dev/dsk/c10t600000E00D000000000201A400020000d0s0
    8.0T 4.3T 3,7T 54% /mnt
    *2. But today I noticed in dmesg such errors: "ufs: [ID 682040 kern.notice] NOTICE: /mnt: out of inodes"*
    -bash-3.00# df -i /mnt
    Filesystem Inodes IUsed IFree IUse% Mounted on
    /dev/dsk/c10t600000E00D000000000201A400020000d0s0
    8753024 8753020 4 100% /mnt
    *3. So, I decided to make file system with new parameters:*
    -bash-3.00# mkfs -m /dev/rdsk/c10t600000E00D000000000201A400020000d0s0
    mkfs -F ufs -o nsect=128,ntrack=48,bsize=8192,fragsize=8192,cgsize=143,free=1,rps=1,nbpi=997778,opt=t,apc=0,gap=0,nrpos=1,maxcontig=128,mtb=y /dev/rdsk/c10t600000E00D000000000201A400020000d0s0 17165172656
    -bash-3.00#
    -bash-3.00# mkfs -F ufs -o nsect=128,ntrack=48,bsize=8192,fragsize=1024,cgsize=143,free=1,rps=1,nbpi=512,opt=t,apc=0,gap=0,nrpos=1,maxcontig=128,mtb=f /dev/rdsk/c10t600000E00D000000000201A400020000d0s0 17165172656
    *3. I've got some warnings about inodes threshold:*
    -bash-3.00# mkfs -F ufs -o nsect=128,ntrack=48,bsize=8192,fragsize=1024,cgsize=143,free=1,rps=1,nbpi=512,opt=t,apc=0,gap=0,nrpos=1,maxcontig=128,mtb=n /dev/rdsk/c10t600000E00D000000000201A400020000d0s0 17165172656
    mkfs: bad value for nbpi: must be at least 1048576 for multi-terabyte, nbpi reset to default 1048576
    Warning: 2128 sector(s) in last cylinder unallocated
    /dev/rdsk/c10t600000E00D000000000201A400020000d0s0: 17165172656 sectors in 2793811 cylinders of 48 tracks, 128 sectors
    8381432.0MB in 19538 cyl groups (143 c/g, 429.00MB/g, 448 i/g)
    super-block backups (for fsck -F ufs -o b=#) at:
    32, 878752, 1757472, 2636192, 3514912, 4393632, 5272352, 6151072, 7029792,
    7908512,
    Initializing cylinder groups:
    super-block backups for last 10 cylinder groups at:
    17157145632, 17158024352, 17158903072, 17159781792, 17160660512, 17161539232,
    17162417952, 17163296672, 17164175392, 17165054112
    *4.And my inodes number didn't change:*
    -bash-3.00# df -i /mnt
    Filesystem Inodes IUsed IFree IUse% Mounted on
    /dev/dsk/c10t600000E00D000000000201A400020000d0s0
    8753024 4 8753020 1% /mnt
    I found http://wesunsolve.net/bugid.php/id/6595253 that is a bug of mkfs without workaround. Is ZFS what I need now?

    Well, to fix the bug you referred to you can apply patch 141444-01 or 141445-01.
    However that bug is just regarding an irrelevant error message from mkfs, it will not fix your problem as such.
    It seems to me like the minimum value for nbpi on a multi-terabyte filesystem is 1048576, hence you won't be able to create a filesystem with more inodes.
    The things to try would be to either create two UFS filesystems, or go with ZFS, which is the future anyway ;-)
    .7/M.

  • Sql query slowness due to rank and columns with null values:

        
    Sql query slowness due to rank and columns with null values:
    I have the following table in database with around 10 millions records:
    Declaration:
    create table PropertyOwners (
    [Key] int not null primary key,
    PropertyKey int not null,    
    BoughtDate DateTime,    
    OwnerKey int null,    
    GroupKey int null   
    go
    [Key] is primary key and combination of PropertyKey, BoughtDate, OwnerKey and GroupKey is unique.
    With the following index:
    CREATE NONCLUSTERED INDEX [IX_PropertyOwners] ON [dbo].[PropertyOwners]    
    [PropertyKey] ASC,   
    [BoughtDate] DESC,   
    [OwnerKey] DESC,   
    [GroupKey] DESC   
    go
    Description of the case:
    For single BoughtDate one property can belong to multiple owners or single group, for single record there can either be OwnerKey or GroupKey but not both so one of them will be null for each record. I am trying to retrieve the data from the table using
    following query for the OwnerKey. If there are same property rows for owners and group at the same time than the rows having OwnerKey with be preferred, that is why I am using "OwnerKey desc" in Rank function.
    declare @ownerKey int = 40000   
    select PropertyKey, BoughtDate, OwnerKey, GroupKey   
    from (    
    select PropertyKey, BoughtDate, OwnerKey, GroupKey,       
    RANK() over (partition by PropertyKey order by BoughtDate desc, OwnerKey desc, GroupKey desc) as [Rank]   
    from PropertyOwners   
    ) as result   
    where result.[Rank]=1 and result.[OwnerKey]=@ownerKey
    It is taking 2-3 seconds to get the records which is too slow, similar time it is taking as I try to get the records using the GroupKey. But when I tried to get the records for the PropertyKey with the same query, it is executing in 10 milliseconds.
    May be the slowness is due to as OwnerKey/GroupKey in the table  can be null and sql server in unable to index it. I have also tried to use the Indexed view to pre ranked them but I can't use it in my query as Rank function is not supported in indexed
    view.
    Please note this table is updated once a day and using Sql Server 2008 R2. Any help will be greatly appreciated.

    create table #result (PropertyKey int not null, BoughtDate datetime, OwnerKey int null, GroupKey int null, [Rank] int not null)Create index idx ON #result(OwnerKey ,rnk)
    insert into #result(PropertyKey, BoughtDate, OwnerKey, GroupKey, [Rank])
    select PropertyKey, BoughtDate, OwnerKey, GroupKey,
    RANK() over (partition by PropertyKey order by BoughtDate desc, OwnerKey desc, GroupKey desc) as [Rank]
    from PropertyOwners
    go
    declare @ownerKey int = 1
    select PropertyKey, BoughtDate, OwnerKey, GroupKey
    from #result as result
    where result.[Rank]=1
    and result.[OwnerKey]=@ownerKey
    go
    Best Regards,Uri Dimant SQL Server MVP,
    http://sqlblog.com/blogs/uri_dimant/
    MS SQL optimization: MS SQL Development and Optimization
    MS SQL Consulting:
    Large scale of database and data cleansing
    Remote DBA Services:
    Improves MS SQL Database Performance
    SQL Server Integration Services:
    Business Intelligence

  • In ColumnChart, how to stop displaying small column for zero value?

    I'm trying to create a ColumnChart that does not display a column when the value is zero. The chart currently looks like this:
    This is a stacked column chart with red values representing Faults and green values representing Throughput. Each hour has an explicit value of zero for Faults. I don't want to see any red lines when the value is zero. Is there any way to accomplish this?
    My code looks like this:
    <mx:Script><![CDATA[
         import mx.collections.ArrayCollection;
         [Bindable]
         public var simpleStats:ArrayCollection = new ArrayCollection([
            {Hour:"0:00", Throughput:0, ThroughputThreshold:2000, Faults:0, MaxResponseTime:450, AvgResponseTime:200, MinResponseTime:180,
                 AuthenticationAcceptance:50, AuthenticationRejection:0, AuthorizationAcceptance:50, AuthorizationRejection:0},    
            {Hour:"1:00", Throughput:0, ThroughputThreshold:2000, Faults:0, MaxResponseTime:450, AvgResponseTime:200, MinResponseTime:180,
                 AuthenticationAcceptance:50, AuthenticationRejection:0, AuthorizationAcceptance:50, AuthorizationRejection:0},    
            {Hour:"2:00", Throughput:0, ThroughputThreshold:2000, Faults:0, MaxResponseTime:450, AvgResponseTime:200, MinResponseTime:180,
                 AuthenticationAcceptance:50, AuthenticationRejection:0, AuthorizationAcceptance:50, AuthorizationRejection:0},    
            {Hour:"3:00", Throughput:5, ThroughputThreshold:2000, Faults:0, MaxResponseTime:450, AvgResponseTime:200, MinResponseTime:180,
                 AuthenticationAcceptance:50, AuthenticationRejection:0, AuthorizationAcceptance:50, AuthorizationRejection:0},    
            {Hour:"4:00", Throughput:0, ThroughputThreshold:2000, Faults:0, MaxResponseTime:450, AvgResponseTime:200, MinResponseTime:180,
                 AuthenticationAcceptance:50, AuthenticationRejection:0, AuthorizationAcceptance:50, AuthorizationRejection:0}
      ]]></mx:Script>
                                     <mx:ColumnChart id="trafficChart"
                                        dataProvider="{simpleStats}"
                                        showDataTips="true" width="500" height="100%" seriesFilters="[]"
                                        type="stacked">
                                        <mx:verticalAxis>
                                            <mx:LinearAxis title="Messages" id="a1"/>
                                        </mx:verticalAxis>                                   
                                        <mx:horizontalAxis>
                                           <mx:CategoryAxis
                                                   id="TrafficTimeAxis"
                                                dataProvider="{simpleStats}"
                                                categoryField="Hour"
                                                />
                                        </mx:horizontalAxis>
                                        <mx:horizontalAxisRenderers>
                                            <mx:AxisRenderer axis="{TrafficTimeAxis}" canDropLabels="true"/>                                       
                                        </mx:horizontalAxisRenderers>                                            
                                        <mx:series>                                                                         
                                           <mx:ColumnSeries
                                                yField="Faults"
                                                displayName="Faults"
                                                fill="{sc2}"
                                                stroke="{s2}"
                                                 minHeight="0">
                                           </mx:ColumnSeries>                                                                                                          
                                           <mx:ColumnSeries
                                                yField="Throughput"
                                                displayName="Throughput"
                                                fill="{sc1}"
                                                stroke="{s1}"
                                                minHeight="0">
                                            </mx:ColumnSeries>                                      
                                        </mx:series>
                                     </mx:ColumnChart>

    Answered my own question!!
    The solution is to set the stroke for the ColumnSeries to {noStroke}, which I defined like this:
        <mx:Stroke id="noStroke" color="0xFFFFFF" weight="0"/>
    Note that if you define {noStroke} without a color property, black is used by default, which means that those columns that do appear show up in black stroke outline.
    If there's a more elegant solution to the problem, please let me know...

  • How to add column access BADI to appraisal template.

    I am not able to see option to add column access BADI in template at Processing  tab--> Further template behaviour.
    Kinldy let me know how to add it.
    I am able to see only following options
    BC         Business Check
    D1         Default Appraiser
    D2         Default Appraisee
    DE         Default Execution Period
    DN         Default Appraisal Document Name
    DP         Default Part Appraiser
    DR         Default Review Date
    DV         Default Validity Period
    HA         Header Data Change
    PM         Maximum Number of Part Appraisers
    Thanks in advance.
    Dhananjy

    Hello Ravikiran,
    Based on the Role and Status of the appraisal form, turn the Column On/Off/Hide..
    sample code...
    case s_header_status-ap_status.
              when c_status_in_preparation.
                availability = 'H'.
                column_owner   = space.
              when  c_status_in_planning.
                case s_body_columns-column_id.
                  when 'OBJ0'.
                    availability   = 'X'.
                    column_owner   = 'X'.
                    exit.
                  when 'Z003'.
                    availability   = 'X'.
                    column_owner   = 'X'.
    Hope it helps...
    Regards,
    Jothi.P

  • How to get LOV column's DISPLAY value in IR report?

    Hi,
    I would like to get LOV column's DISPLAY value to trigger another javascript function. Could you please provide any hint?
    For example, there is one IR report with 2 column, DEPT_NAME & SHOW.
    The DEPT_NAME type is LOV .
    The SHOW column is one link image, when clicking it , I would like to fire a javascript to get DEPT_NAME 's display value.
    I know I can use #DEPT_NAME# to get the return value of LOV. However, how to get current row's DEPT_NAME's display value?
    Thanks in advance.
    Ray
    Edited by: 最爱用中文 on 2013-4-11 下午4:14

    Is there any way you can create a table join, instead of using LOV as your display type?
    That way you can select the name of the department in your IR query, display it as text, and then easily reference it as #DEPARTMENT_NAME#, for example.

  • OPC fieldpoint VT_BOOL gives "Bad Value" in Citect

    We run Citects OPC-client and tries to communicate with the Fieldpoint OPC-server. Analog values AI/AO work all fine but DIO-550 I/O-values do not. The data type is VT_BOOL. The error message is "Bad value". The client can be configured to skip "bad value"-error. That solves DO but not DI.
    Testing Macs OPC-client gives same problem.
    Our Fieldpoint OPC-server works fine with NI Server Explorer and the Citect and Macs clients works fine with other OPC-servers! 
    Please someone give us a hint of what to do.
    /Dan Larsson, Damill AB, Sweden

    We run Citects OPC-client and tries to communicate with the Fieldpoint OPC-server. Analog values AI/AO work all fine but DIO-550 I/O-values do not. The data type is VT_BOOL. The error message is "Bad value". The client can be configured to skip "bad value"-error. That solves DO but not DI.
    Testing Macs OPC-client gives same problem.
    Our Fieldpoint OPC-server works fine with NI Server Explorer and the Citect and Macs clients works fine with other OPC-servers! 
    Please someone give us a hint of what to do.
    /Dan Larsson, Damill AB, Sweden

  • Adding virtual column: ORA-12899: value too large for column

    I'm using Oracle 11g, Win7 OS, SQL Developer
    I'm trying to add virtual column to my test table, but getting ORA-12899: value too large for column error. Below are the details.
    Can someone help me in this?
    CREATE TABLE test_reg_exp
    (col1 VARCHAR2(100));
    INSERT INTO test_reg_exp (col1) VALUES ('ABCD_EFGH');
    INSERT INTO test_reg_exp (col1) VALUES ('ABCDE_ABC');
    INSERT INTO test_reg_exp (col1) VALUES ('WXYZ_ABCD');
    INSERT INTO test_reg_exp (col1) VALUES ('ABCDE_PQRS');
    INSERT INTO test_reg_exp (col1) VALUES ('ABCD_WXYZ');
    ALTER TABLE test_reg_exp
    ADD (col2 VARCHAR2(100) GENERATED ALWAYS AS (REGEXP_REPLACE (col1, '^ABCD[A-Z]*_')));
    SQL Error: ORA-12899: value too large for column "COL2" (actual: 100, maximum: 400)
    12899. 00000 -  "value too large for column %s (actual: %s, maximum: %s)"
    *Cause:    An attempt was made to insert or update a column with a value
               which is too wide for the width of the destination column.
               The name of the column is given, along with the actual width
               of the value, and the maximum allowed width of the column.
               Note that widths are reported in characters if character length
               semantics are in effect for the column, otherwise widths are
               reported in bytes.
    *Action:   Examine the SQL statement for correctness.  Check source
               and destination column data types.
               Either make the destination column wider, or use a subset
               of the source column (i.e. use substring).When I try to select, I'm getting correct results:
    SELECT col1, (REGEXP_REPLACE (col1, '^ABCD[A-Z]*_'))
    FROM test_reg_exp;Thanks.

    Yes RP, it working if you give col2 size >=400.
    @Northwest - Could you please test the same w/o having a regex clause in col2?
    I doubt on the usage of a REGEX in this dynamic col case.
    Refer this (might help) -- http://www.oracle-base.com/articles/11g/virtual-columns-11gr1.php
    Below snippet from above link.... see if this helps...
    >
    Notes and restrictions on virtual columns include:
    Indexes defined against virtual columns are equivalent to function-based indexes.
    Virtual columns can be referenced in the WHERE clause of updates and deletes, but they cannot be manipulated by DML.
    Tables containing virtual columns can still be eligible for result caching.
    Functions in expressions must be deterministic at the time of table creation, but can subsequently be recompiled and made non-deterministic without invalidating the virtual column. In such cases the following steps must be taken after the function is recompiled:
    Constraint on the virtual column must be disabled and re-enabled.
    Indexes on the virtual column must be rebuilt.
    Materialized views that access the virtual column must be fully refreshed.
    The result cache must be flushed if cached queries have accessed the virtual column.
    Table statistics must be regathered.
    Virtual columns are not supported for index-organized, external, object, cluster, or temporary tables.
    The expression used in the virtual column definition has the following restrictions:
    It cannot refer to another virtual column by name.
    It can only refer to columns defined in the same table.
    If it refers to a deterministic user-defined function, it cannot be used as a partitioning key column.
    The output of the expression must be a scalar value. It cannot return an Oracle supplied datatype, a user-defined type, or LOB or LONG RAW.
    >
    Edited by: ranit B on Oct 16, 2012 11:48 PM
    Edited by: ranit B on Oct 16, 2012 11:54 PM

  • What is the key column name and value column name in JDBC Adapter parameter

    Hi
    Can any one please tell me what is the Key Column Name and Key Column Value in JDBC adatper parameters. If i dont mention those parameters i am getting the following error
    <b> Value missing for mandatory configuration attribute tableEOColumnNameId</b>
    Please help me
    Best Regards
    Ravi Shankar B

    Hi
    I am doing DataBase Lookup in XI
    First i have created a  Table in Database( CheckUser) which has two fields UserName and PhoneNumber and then i have created
    I have created one Communication Channel For Reciever Adapter .
    I have given the parameters like this
    JDBC Driver : com.microsoft.jdbc.sqlserver.SQLServerDriver
    Connection : jdbc:microsoft:sqlserver://10.7.1.43:1433;DatabaseName=Ravi;
    UserName.... sa
    password.... sa
    persistence : Database
    Database Table Name : CheckUser
    Key column name and Value column name i left blank and activated
    and then
    I have created
    Data Types : Source ...... UserName
                      Destination.... PhoneNumber
    Message Types
    Message Interfaces
    In Message Mapping
                  I have created one User Defined function DBProcessing_SpecialAPI().This method will get the data from the database....
    In this function i have written the following code
       //write your code here
    String query = " ";
    Channel channel = null;
    DataBaseAccessor accessor = null;
    DataBaseResult resultSet = null;
    query = "select Password from CheckUser where UserName = ' " +UserName[0]+ " '  ";
    try {
         channel = LookupService.getChannel("Ravi","CC_JDBC");
         accessor = LookupService.getDataBaseAccessor(channel);
         resultSet = accessor.execute(query);
         for(Iterator rows = resultSet.getRows();rows.hasNext();){
              Map  rowMap = (Map)rows.next();
              result.addValue((String)rowMap.get("Password"));
    catch(Exception e){
         result.addValue(e.getMessage());
    finally{
         try{
              if(accessor != null)
                   accessor.close();
         }catch(Exception e){
              result.addValue(e.getMessage());
    And the i have mapped like this
    UserName -
    > DBProcessing_SpecialAPI----
    >PhoneNumber
    when i am testing this mapping i am getting the following error
    <?xml version="1.0" encoding="UTF-8"?>
    <ns0:Dest_JDBC_MT xmlns:ns0="http://filetofilescenario.com/ilg"><phoneNumber>Plain exception:Problem when calling an adapter by using communication channel CC_JDBC (Party: , Service: Ravi, Object ID: c360bc139a403293afbc49d5c46e4478) Check whether the communication channel exists in the Integration Directory; also check the cache notifications for the instance Integration Server (central Adapter-Engine) Channel object with Id Channel:c360bc139a403293afbc49d5c46e4478 not available in CPA Cache.
    com.sap.aii.mapping.lookup.LookupException: Problem when calling an adapter by using communication channel CC_JDBC (Party: , Service: Ravi, Object ID: c360bc139a403293afbc49d5c46e4478) Check whether the communication channel exists in the Integration Directory; also check the cache notifications for the instance Integration Server (central Adapter-Engine) Channel object with Id Channel:c360bc139a403293afbc49d5c46e4478 not available in CPA Cache.
         at com.sap.aii.ibrun.server.lookup.AdapterProxyLocal.<init>(AdapterProxyLocal.java:61)
         at com.sap.aii.ibrun.server.lookup.SystemAccessorInternal.getProxy(SystemAccessorInternal.java:98)
         at com.sap.aii.ibrun.server.lookup.SystemAccessorInternal.<init>(SystemAccessorInternal.java:38)
         at com.sap.aii.ibrun.server.lookup.SystemAccessorHmiServer.getConnection(SystemAccessorHmiServer.java:270)
         at com.sap.aii.ibrun.server.lookup.SystemAccessorHmiServer.process(SystemAccessorHmiServer.java:70)
         at com.sap.aii.utilxi.hmis.server.HmisServiceImpl.invokeMethod(HmisServiceImpl.java:169)
         at com.sap.aii.utilxi.hmis.server.HmisServer.process(HmisServer.java:178)
         at com.sap.aii.utilxi.hmis.web.HmisServletImpl.processRequestByHmiServer(HmisServletImpl.java:296)
         at com.sap.aii.utilxi.hmis.web.HmisServletImpl.processRequestByHmiServer(HmisServletImpl.java:211)
         at com.sap.aii.utilxi.hmis.web.workers.HmisInternalClient.doWork(HmisInternalClient.java:70)
         at com.sap.aii.utilxi.hmis.web.HmisServletImpl.doWork(HmisServletImpl.java:496)
         at com.sap.aii.utilxi.hmis.web.HmisServletImpl.doPost(HmisServletImpl.java:634)
         at javax.servlet.http.HttpServlet.service(HttpServlet.java:760)
         at javax.servlet.http.HttpServlet.service(HttpServlet.java:853)
         at com.sap.engine.services.servlets_jsp.server.HttpHandlerImpl.runServlet(HttpHandlerImpl.java:390)
         at com.sap.engine.services.servlets_jsp.server.HttpHandlerImpl.handleRequest(HttpHandlerImpl.java:264)
         at com.sap.engine.services.httpserver.server.RequestAnalizer.startServlet(RequestAnalizer.java:347)
         at com.sap.engine.services.httpserver.server.RequestAnalizer.startServlet(RequestAnalizer.java:325)
         at com.sap.engine.services.httpserver.server.RequestAnalizer.invokeWebContainer(RequestAnalizer.java:887)
         at com.sap.engine.services.httpserver.server.RequestAnalizer.handle(RequestAnalizer.java:241)
         at com.sap.engine.services.httpserver.server.Client.handle(Client.java:92)
         at com.sap.engine.services.httpserver.server.Processor.request(Processor.java:148)
         at com.sap.engine.core.service630.context.cluster.session.ApplicationSessionMessageListener.process(ApplicationSessionMessageListener.java:33)
         at com.sap.engine.core.cluster.impl6.session.MessageRunner.run(MessageRunner.java:41)
         at com.sap.engine.core.thread.impl3.ActionObject.run(ActionObject.java:37)
         at java.security.AccessController.doPrivileged(Native Method)
         at com.sap.engine.core.thread.impl3.SingleThread.execute(SingleThread.java:100)
         at com.sap.engine.core.thread.impl3.SingleThread.run(SingleThread.java:170)
    </phoneNumber></ns0:Dest_JDBC_MT>
    In RWB i have checked the status of JDBC driver its showing the following error
    <b>Value missing for mandatory configuration attribute tableEOColumnNameId</b>
    Best Regards
    Ravi Shankar B
    Message was edited by:
            RaviShankar B

  • Update column with ROW_NUMBER() value

    I have a column that I would like to update with a ROW_NUMBER() value partitioned by a specific column.
    CREATE TABLE ##tmp (id INT NULL,
    value1 VARCHAR(10),
    value2 VARCHAR(10))
    INSERT INTO ##tmp (value1,
    value2)
    VALUES ('A', 'asdfasdf'),
    ('A', 'asdf'),
    ('A', 'VC'),
    ('B', 'aasdf'),
    ('C', 'sdfgs'),
    ('C', 'xdfbhsdty'),
    ('C', '23sdgfg'),
    ('C', '234')
    -- update the ID column with the values below
    SELECT ROW_NUMBER() OVER (PARTITION BY Value1 ORDER BY Value2),
    Value1,
    Value2
    FROM ##tmp
    I have 14 million records.  Can someone explain the best way to do this?  Does the ORDER BY destroy performance for this many records?
    Thanks!

    CREATE TABLE ##tmp (id INT NULL,
    value1 VARCHAR(10),
    value2 VARCHAR(10))
    INSERT INTO ##tmp (value1,
    value2)
    VALUES ('A', 'asdfasdf'),
    ('A', 'asdf'),
    ('A', 'VC'),
    ('B', 'aasdf'),
    ('C', 'sdfgs'),
    ('C', 'xdfbhsdty'),
    ('C', '23sdgfg'),
    ('C', '234')
    -- update the ID column with the values below
    ; with cte as (
    SELECT ROW_NUMBER() OVER (PARTITION BY Value1 ORDER BY Value2) as RowNumber,
    Value1,
    Value2
    FROM ##tmp
    update cte
    set Id = RowNumber;
    Russel Loski, MCT, MCSE Data Platform/Business Intelligence. Twitter: @sqlmovers; blog: www.sqlmovers.com

Maybe you are looking for

  • Automatic event creater in iCal

    I work at a school that has a seven day cycle: A through G days. I am trying to create a workflow that will automatically create events (my classes) for each day in my cycle. For example, this Thursday will be an A Day, and I will have one class from

  • Images, text, etc. display poorly, they are a little fuzzy. IE and Chrome display the imags as they are, very crisp... Why?

    I notice that the image quality in Firefox is not as good as other browsers. I have two monitors and have moved the browser into each so it's not my monitor. What's the fix?

  • Download Photo Albums on iWeb?

    The photos we have on our iWeb are able to be downloaded individually by people. I would like to be able to have the photo albums be downloaded as well, so that people do not have to download each photo individually. Is this possible and, if so, how

  • Problem with album picture on iPod with iTunes 10.2.2.14

    I'm using iTunes 10.2.2.14 on windows xp (service pack3) with iPod classic 80Gb. In iTunes I have 14.000 songs. Most of them with album illustration. In iTunes these illustrations are available. On my iPod I can se the covers On my iPod Classic in [m

  • WRT54GC - Orb app not port forwarding

    When trying to configure an app for port forwarding sometimes it will time out. In the App and Gaming section I will specify the app name, the start and end port number and finally enable it. When I click save settings it will just wait and wait and