Phpamf not loading sql-Longtext field

I'm loading data from SQL using phpamf service.
I have employees table with Id, Name and Image.
Im storing the image as Base64-string format and that is why need Longtext field in the table.
When I drag the service getAllEmployees to a DataGrid, Id and Name are loaded, but Image-string is blank.
Trying to create image from the base64-string fails too (no data is fetched from db for this field).
Please tell me if this is a known issue.
Give ideas for solving it or better way to store images in database (give exaplme please).
Thanks in advance!

Thanks for the response, but that didn't seem to fix the issue for me:
SQL*Loader-350: Syntax error at line 24.
Expecting quoted string or hex identifier, found "0".
COL_D DECIMAL EXTERNAL NULLIF COL_D=0 ENCLOSED
^

Similar Messages

  • Auto Increment Not Working - SQL Identity field

    I am new to Visual Basic & SQL server but have some experience in Access & VBA. But this is a steep learning curve!
    I am banging my head against a wall with a problem..
    All of my tables in this database are in the dataset. The main one being tblItems with a PK 'ITEMID'
    I have 2 forms - the first one is used to lookup an item the second displays the item's full details.
    On the first form (lookup) I have a 'Add New' button which launches the second form with the code - frmProductDetail.VItemsBindingSource.AddNew()
    This opens the form with empty boxes as expected. I have a 'Save' button on the second form with the following code -
            Dim row As SASHItemsDataSet.tblItemsRow
            row = SASHItemsDataSet.tblItems.NewRow
            With row
                .ITEMCODE = txtItemCode.Text
                .ITEMDESCRIPTION = txtItemDescription.Text
                .CATEGORY = cmbItemCategory.SelectedValue
                .PURCHCOST = txtPurchCost.Text
                .SELLCOST = txtSellPrice.Text
                .UNIT = cmbUOM.SelectedValue
                .VATID = cmbVAT.SelectedValue
                .WHLOCATION = cmbWHLoc.SelectedValue
            End With
            SASHItemsDataSet.tblItems.Rows.Add()
            Try
                Me.Validate()
                Me.VItemsBindingSource.EndEdit()
            Catch ex As Exception
                MsgBox(ex.Message, MsgBoxStyle.Critical + MsgBoxStyle.OkOnly, "UPDATE FAILED")
            End Try
    My problem is I get the error msg box with the following error 'Column 'ITEMID' does not allow nulls'
    This field is set as a auto incrementing identity field with all the correct settings shown in Visual Studio so it shouldn't be coming back as null.
    I have Googled for hours & tried all sorts with no luck..
    I have clearly gone wrong somewhere but I can't work out where... any help appreciated!
    James

    This is the code on frmProductLookup that opens the form...
            frmProductDetail.Show()
            frmProductDetail.VItemsBindingSource.AddNew()
    This is the code on frmProductDetail_Load...
            'TODO: This line of code loads data into the 'SASHItemsDataSet.tblVAT' table. You can move, or remove it, as needed.
            Me.TblVATTableAdapter.Fill(Me.SASHItemsDataSet.tblVAT)
            'TODO: This line of code loads data into the 'SASHItemsDataSet.tblWarehouseLocations' table. You can move, or remove it, as needed.
            Me.TblWarehouseLocationsTableAdapter.Fill(Me.SASHItemsDataSet.tblWarehouseLocations)
            'TODO: This line of code loads data into the 'SASHItemsDataSet.tblStockUnits' table. You can move, or remove it, as needed.
            Me.TblStockUnitsTableAdapter.Fill(Me.SASHItemsDataSet.tblStockUnits)
            'TODO: This line of code loads data into the 'SASHItemsDataSet.tblItemCategory' table. You can move, or remove it, as needed.
            Me.TblItemCategoryTableAdapter.Fill(Me.SASHItemsDataSet.tblItemCategory)
            'TODO: This line of code loads data into the 'SASHItemsDataSet.vItems' table. You can move, or remove it, as needed.
            'Me.VItemsTableAdapter.Fill(Me.SASHItemsDataSet.vItems)
            Me.VItemsTableAdapter.Fill(Me.SASHItemsDataSet.vItems)
            Me.VItemsBindingSource.Position = Me.VItemsBindingSource.Find("ITEMID", _passedText)
            Me.txtWKSName.Text = Environment.MachineName
            Me.txtSellPrice.Text = FormatCurrency(Me.txtSellPrice.Text)
            Me.txtPurchCost.Text = FormatCurrency(Me.txtPurchCost.Text)
            Me.txtPriceIVAT.Text = FormatCurrency(Me.txtSellPrice.Text + (Me.txtSellPrice.Text * 0.2))
    On the tblItemTableAdapter this is the Command Text for the Insert Command...
    INSERT INTO tblItems
                             (ITEMCODE, ITEMDESCRIPTION, UNIT, WHLOCATION, VATID, PREFSUPPLIER, CATEGORY, PURCHCOST, SELLCOST, INACTIVE)
    VALUES        (@ITEMCODE,@ITEMDESCRIPTION,@UNIT,@WHLOCATION,@VATID,@PREFSUPPLIER,@CATEGORY,@PURCHCOST,@SELLCOST,@INACTIVE)
    SQL Server Management Studio clearly shows the column as an Identity column. If I add a row through SSMS it does create an PK identity automatically. I have also dropped the ITEMID column & recreated through SSMS which has no effect.
    I am now considering creating a separate form just for adding an item but I have managed it in Access but am just struggling with VB!
    Thanks,
    James

  • Data not loading for new Fields InfoObjects from DSO to InfoCube

    Hi Gurus -
                   I have a DataSource that is providing data to existing DSO then to Infocube. My client asked me to added couple fields to DataSource and get the data to DSO and then DSO to InfoCube.
    Here is the Old scenarion:   DataSource -> DSO -> InfoCube.
    Here is the New Scenation:
         DataSource ( Added new fields) -> DSO (added new InfoObjects for Corresponding DataSource FIelds) -> InfoCube( Added new InfoObjects to mapped from DSO)
    I added the new fields to DataSource, added the corresponding InfoObjects to DSO and InfoCube.
    I successfully loaded data from DataSource to DSO. Data is populating for the new Fields/InfoObjects in DSO.
    But when I load data from DSO to InfoCube, I don't see any data for New Fields InfoObjects in the InfoCube.
    Data from DSO to InfoCube is loading fine for the Old InfoObjects Fields but not for the New InfoObjects I added in InfoCube.
    -Sonali

    Hi,
    Why dont u debug the load through DTP debugging and check what happens to the source field and target field once it passes through the transformation. You can easily trace back where the fields are becoming blank.
    The loads which you have mentioned earlier did it have values in Added Records/Transferred Records column for the cube.
    Regards,
    Mani

  • SQL Loader Not Loading Data

    According to sql loaders log file, I was successful in loading my data into a table. However the table does not show any values loaded - all empy.
    Here's my table:
    SQL> desc patients
    Name Null? Type
    PATIENT_ID NOT NULL NUMBER(38)
    MEDSUBSCRIBERID VARCHAR2(200)
    MEDMEMBERID VARCHAR2(50)
    MEDDEPSEQDESC VARCHAR2(50)
    MEDMEMBERGENDER VARCHAR2(50)
    MEDMEMBERDOB DATE
    MEDDEPSEQCODE VARCHAR2(50)
    NETWORKID NUMBER(38)
    NETWORKCODE VARCHAR2(50)
    PATIENTAGE NUMBER(38)
    PATIENTNAME VARCHAR2(50)
    PATIENTLASTNAME VARCHAR2(50)
    PATIENTSEX VARCHAR2(50)
    PATIENTSSN VARCHAR2(50)
    INTMEDSUBSCRIBERID VARCHAR2(50)
    DATE_VALUE_ID NUMBER(38)
    DOB VARCHAR2(50)
    AGE_GROUP_ID NUMBER(38)
    ORGANIZATIONNETWORK_ID NUMBER(38)
    ORGANIZATION_ID NUMBER(38)
    ORGANIZATIONCODE VARCHAR2(50)
    NETWORKID_HOLD NUMBER(38)
    PATIENTRELATION VARCHAR2(50)
    Here's my SQL Loader Control File:
    LOAD DATA
    INFILE 'patients.dat'
    APPEND
    INTO TABLE PATIENTS
    FIELDS TERMINATED BY '|'
    TRAILING NULLCOLS
    PATIENT_ID INTEGER(8) NULLIF(PATIENT_ID=BLANKS)
    , MEDSUBSCRIBERID CHAR NULLIF(MEDSUBSCRIBERID=BLANKS)
    , MEDMEMBERID CHAR NULLIF(MEDMEMBERID=BLANKS)
    , MEDDEPSEQDESC CHAR NULLIF(MEDDEPSEQDESC=BLANKS)
    , MEDMEMBERGENDER CHAR NULLIF(MEDMEMBERGENDER=BLANKS)
    , MEDMEMBERDOB DATE "YYYY-MM-DD HH24:MI:SS" NULLIF(MEDMEMBERDOB=BLANKS)
    , MEDDEPSEQCODE CHAR NULLIF(MEDDEPSEQCODE=BLANKS)
    , NETWORKID INTEGER(8) NULLIF(NETWORKID=BLANKS)
    , NETWORKCODE CHAR NULLIF(NETWORKCODE=BLANKS)
    , PATIENTAGE INTEGER(8) NULLIF(PATIENTAGE=BLANKS)
    , PATIENTNAME CHAR NULLIF(PATIENTNAME=BLANKS)
    , PATIENTLASTNAME CHAR NULLIF(PATIENTLASTNAME=BLANKS)
    , PATIENTSEX CHAR NULLIF(PATIENTSEX=BLANKS)
    , PATIENTSSN CHAR NULLIF(PATIENTSSN=BLANKS)
    , INTMEDSUBSCRIBERID CHAR NULLIF(INTMEDSUBSCRIBERID=BLANKS)
    , DATE_VALUE_ID INTEGER(8) NULLIF(DATE_VALUE_ID=BLANKS)
    , DOB CHAR NULLIF(DOB=BLANKS)
    , AGE_GROUP_ID INTEGER(8) NULLIF(AGE_GROUP_ID=BLANKS)
    , ORGANIZATIONNETWORK_ID INTEGER(8) NULLIF(ORGANIZATIONNETWORK_ID=BLANKS)
    , ORGANIZATION_ID INTEGER(8) NULLIF(ORGANIZATION_ID=BLANKS)
    , ORGANIZATIONCODE CHAR NULLIF(ORGANIZATIONCODE=BLANKS)
    , NETWORKID_HOLD INTEGER(8) NULLIF(NETWORKID_HOLD=BLANKS)
    , PATIENTRELATION CHAR NULLIF(PATIENTRELATION=BLANKS)
    Here's data file that I'm trying to load:
    2|001321666|00132156801||M|1944-01-22 00:00:00||42|7215||Donald|Duck||||52616|19440666|4|42|2|WELLS|42|E
    3|001321999|00132156802||F|1951-09-01 00:00:00||42|7215||Mickey|Mouse||||55395|19510999|3|42|2|WELLS|42|W
    4|001363888|00136366201||M|1955-05-22 00:00:00||42|7215||Daffy|Duck||||56754|19550777|3|42|2|WELLS|42|E
    And, my log file after executing sql loader successfully but the data does not load:
    Table PATIENTS:
    3 Rows successfully loaded.
    0 Rows not loaded due to data errors.
    0 Rows not loaded because all WHEN clauses were failed.
    0 Rows not loaded because all fields were null.
    Space allocated for bind array: 251776 bytes(64 rows)
    Read buffer bytes: 1048576
    Total logical records skipped: 0
    Total logical records read: 3
    Total logical records rejected: 0
    Total logical records discarded: 0
    Run began on Wed Aug 20 11:51:51 2008
    Run ended on Wed Aug 20 11:51:57 2008
    Elapsed time was: 00:00:05.47
    CPU time was: 00:00:00.02

    Very strange.
    Did you connect as the same user as SQL*Loader when you checked the table?
    Is it possible that someone else deleted the data?

  • SQl loader not loading records

    I have my control file like this
    options (skip=1)
    LOAD DATA
    INFILE xxx.csv
    into table xxx
    TRUNCATE
    FIELDS TERMINATED BY ',' optionally enclosed by '"'
         RECORD_STATUS,
         ITEM_NUMBER,
    Sql loader not loading records and giving error like .......
    Commit point reached - logical record count 14
    Commit point reached - logical record count 26
    Commit point reached - logical record count 84
    Commit point reached - logical record count 92
    and successfully loaded only 41 records among 420 records
    Plz help me

    HI Phiri,
    Thx for your reply.Here is the log file.
    SQL*Loader: Release 8.0.6.3.0 - Production on Wed May 12 21:26:30 2010
    (c) Copyright 1999 Oracle Corporation. All rights reserved.
    Control File: saba_price_break_allcur_test.ctl
    Data File: saba_price_break_allcur_test.csv
    Bad File: saba_price_break_allcur_test.bad
    Discard File: none specified
    (Allow all discards)
    Number to load: ALL
    Number to skip: 1
    Errors allowed: 50
    Bind array: 64 rows, maximum of 65536 bytes
    Continuation: none specified
    Path used: Conventional
    Table SABA_PRICE_BREAK_ALLCUR_TEST, loaded from every logical record.
    Insert option in effect for this table: TRUNCATE
    Column Name Position Len Term Encl Datatype
    RECORD_STATUS FIRST * , O(") CHARACTER
    ITEM_NUMBER NEXT * , O(") CHARACTER
    PA1 NEXT * , O(") CHARACTER
    PA2 NEXT * , O(") CHARACTER
    UOM_CODE NEXT * , O(") CHARACTER
    RANGE_PRICING NEXT * , O(") CHARACTER
    RANGE_FROM NEXT * , O(") CHARACTER
    RANGE_TO NEXT * , O(") CHARACTER
    PRICING_ATTRIBUTE NEXT * , O(") CHARACTER
    PRICING_METHOD NEXT * , O(") CHARACTER
    PRICE_BREAK_LINE_NO NEXT * , O(") CHARACTER
    TEMPLATE_NAME NEXT * , O(") CHARACTER
    ITEM_DESC NEXT * , O(") CHARACTER
    PRICE_USD NEXT * , O(") CHARACTER
    PRICE_EUR NEXT * , O(") CHARACTER
    PRICE_GBP NEXT * , O(") CHARACTER
    PRICE_JPY NEXT * , O(") CHARACTER
    GL_ACCOUNT NEXT * , O(") CHARACTER
    LONG_DESC NEXT * , O(") CHARACTER
    STATUS NEXT * , O(") CHARACTER
    MESSAGE NEXT * , O(") CHARACTER
    Record 12: Rejected - Error on table SABA_PRICE_BREAK_ALLCUR_TEST, column LONG_DESC.
    Field in data file exceeds maximum length
    Record 13: Rejected - Error on table SABA_PRICE_BREAK_ALLCUR_TEST, column LONG_DESC.
    Field in data file exceeds maximum length
    Record 27: Rejected - Error on table SABA_PRICE_BREAK_ALLCUR_TEST, column LONG_DESC.
    Field in data file exceeds maximum length
    Record 28: Rejected - Error on table SABA_PRICE_BREAK_ALLCUR_TEST, column LONG_DESC.
    Field in data file exceeds maximum length
    Record 29: Rejected - Error on table SABA_PRICE_BREAK_ALLCUR_TEST, column LONG_DESC.
    Field in data file exceeds maximum length
    Record 30: Rejected - Error on table SABA_PRICE_BREAK_ALLCUR_TEST, column LONG_DESC.
    Field in data file exceeds maximum length
    Record 31: Rejected - Error on table SABA_PRICE_BREAK_ALLCUR_TEST, column LONG_DESC.
    Field in data file exceeds maximum length
    Record 32: Rejected - Error on table SABA_PRICE_BREAK_ALLCUR_TEST, column LONG_DESC.
    Field in data file exceeds maximum length
    Record 33: Rejected - Error on table SABA_PRICE_BREAK_ALLCUR_TEST, column LONG_DESC.
    Field in data file exceeds maximum length
    Record 34: Rejected - Error on table SABA_PRICE_BREAK_ALLCUR_TEST, column LONG_DESC.
    Field in data file exceeds maximum length
    Record 35: Rejected - Error on table SABA_PRICE_BREAK_ALLCUR_TEST, column LONG_DESC.
    Field in data file exceeds maximum length
    Record 36: Rejected - Error on table SABA_PRICE_BREAK_ALLCUR_TEST, column LONG_DESC.
    Field in data file exceeds maximum length
    Record 37: Rejected - Error on table SABA_PRICE_BREAK_ALLCUR_TEST, column LONG_DESC.
    Field in data file exceeds maximum length
    Record 38: Rejected - Error on table SABA_PRICE_BREAK_ALLCUR_TEST, column LONG_DESC.
    Field in data file exceeds maximum length
    Record 39: Rejected - Error on table SABA_PRICE_BREAK_ALLCUR_TEST, column LONG_DESC.
    Field in data file exceeds maximum length
    Record 40: Rejected - Error on table SABA_PRICE_BREAK_ALLCUR_TEST, column LONG_DESC.
    Field in data file exceeds maximum length
    Record 41: Rejected - Error on table SABA_PRICE_BREAK_ALLCUR_TEST, column LONG_DESC.
    Field in data file exceeds maximum length
    Record 42: Rejected - Error on table SABA_PRICE_BREAK_ALLCUR_TEST, column LONG_DESC.
    Field in data file exceeds maximum length
    Record 43: Rejected - Error on table SABA_PRICE_BREAK_ALLCUR_TEST, column LONG_DESC.
    Field in data file exceeds maximum length
    Record 44: Rejected - Error on table SABA_PRICE_BREAK_ALLCUR_TEST, column LONG_DESC.
    Field in data file exceeds maximum length
    Record 45: Rejected - Error on table SABA_PRICE_BREAK_ALLCUR_TEST, column LONG_DESC.
    Field in data file exceeds maximum length
    Record 46: Rejected - Error on table SABA_PRICE_BREAK_ALLCUR_TEST, column LONG_DESC.
    Field in data file exceeds maximum length
    Record 47: Rejected - Error on table SABA_PRICE_BREAK_ALLCUR_TEST, column LONG_DESC.
    Field in data file exceeds maximum length
    Record 48: Rejected - Error on table SABA_PRICE_BREAK_ALLCUR_TEST, column LONG_DESC.
    Field in data file exceeds maximum length
    Record 49: Rejected - Error on table SABA_PRICE_BREAK_ALLCUR_TEST, column LONG_DESC.
    Field in data file exceeds maximum length
    Record 50: Rejected - Error on table SABA_PRICE_BREAK_ALLCUR_TEST, column LONG_DESC.
    Field in data file exceeds maximum length
    Record 51: Rejected - Error on table SABA_PRICE_BREAK_ALLCUR_TEST, column LONG_DESC.
    Field in data file exceeds maximum length
    Record 52: Rejected - Error on table SABA_PRICE_BREAK_ALLCUR_TEST, column LONG_DESC.
    Field in data file exceeds maximum length
    Record 53: Rejected - Error on table SABA_PRICE_BREAK_ALLCUR_TEST, column LONG_DESC.
    Field in data file exceeds maximum length
    Record 54: Rejected - Error on table SABA_PRICE_BREAK_ALLCUR_TEST, column LONG_DESC.
    Field in data file exceeds maximum length
    Record 55: Rejected - Error on table SABA_PRICE_BREAK_ALLCUR_TEST, column LONG_DESC.
    Field in data file exceeds maximum length
    Record 56: Rejected - Error on table SABA_PRICE_BREAK_ALLCUR_TEST, column LONG_DESC.
    Field in data file exceeds maximum length
    Record 57: Rejected - Error on table SABA_PRICE_BREAK_ALLCUR_TEST, column LONG_DESC.
    Field in data file exceeds maximum length
    Record 58: Rejected - Error on table SABA_PRICE_BREAK_ALLCUR_TEST, column LONG_DESC.
    Field in data file exceeds maximum length
    Record 59: Rejected - Error on table SABA_PRICE_BREAK_ALLCUR_TEST, column LONG_DESC.
    Field in data file exceeds maximum length
    Record 60: Rejected - Error on table SABA_PRICE_BREAK_ALLCUR_TEST, column LONG_DESC.
    Field in data file exceeds maximum length
    Record 61: Rejected - Error on table SABA_PRICE_BREAK_ALLCUR_TEST, column LONG_DESC.
    Field in data file exceeds maximum length
    Record 62: Rejected - Error on table SABA_PRICE_BREAK_ALLCUR_TEST, column LONG_DESC.
    Field in data file exceeds maximum length
    Record 63: Rejected - Error on table SABA_PRICE_BREAK_ALLCUR_TEST, column LONG_DESC.
    Field in data file exceeds maximum length
    Record 64: Rejected - Error on table SABA_PRICE_BREAK_ALLCUR_TEST, column LONG_DESC.
    Field in data file exceeds maximum length
    Record 65: Rejected - Error on table SABA_PRICE_BREAK_ALLCUR_TEST, column LONG_DESC.
    Field in data file exceeds maximum length
    Record 66: Rejected - Error on table SABA_PRICE_BREAK_ALLCUR_TEST, column LONG_DESC.
    Field in data file exceeds maximum length
    Record 67: Rejected - Error on table SABA_PRICE_BREAK_ALLCUR_TEST, column LONG_DESC.
    Field in data file exceeds maximum length
    Record 68: Rejected - Error on table SABA_PRICE_BREAK_ALLCUR_TEST, column LONG_DESC.
    Field in data file exceeds maximum length
    Record 69: Rejected - Error on table SABA_PRICE_BREAK_ALLCUR_TEST, column LONG_DESC.
    Field in data file exceeds maximum length
    Record 70: Rejected - Error on table SABA_PRICE_BREAK_ALLCUR_TEST, column LONG_DESC.
    Field in data file exceeds maximum length
    Record 73: Rejected - Error on table SABA_PRICE_BREAK_ALLCUR_TEST, column LONG_DESC.
    Field in data file exceeds maximum length
    Record 74: Rejected - Error on table SABA_PRICE_BREAK_ALLCUR_TEST, column LONG_DESC.
    Field in data file exceeds maximum length
    Record 87: Rejected - Error on table SABA_PRICE_BREAK_ALLCUR_TEST, column LONG_DESC.
    Field in data file exceeds maximum length
    Record 91: Rejected - Error on table SABA_PRICE_BREAK_ALLCUR_TEST, column LONG_DESC.
    Field in data file exceeds maximum length
    Record 92: Rejected - Error on table SABA_PRICE_BREAK_ALLCUR_TEST, column LONG_DESC.
    Field in data file exceeds maximum length
    MAXIMUM ERROR COUNT EXCEEDED - Above statistics reflect partial run.
    Table SABA_PRICE_BREAK_ALLCUR_TEST:
    41 Rows successfully loaded.
    51 Rows not loaded due to data errors.
    0 Rows not loaded because all WHEN clauses were failed.
    0 Rows not loaded because all fields were null.
    Space allocated for bind array: 65016 bytes(12 rows)
    Space allocated for memory besides bind array: 0 bytes
    Total logical records skipped: 1
    Total logical records read: 92
    Total logical records rejected: 51
    Total logical records discarded: 0
    Run began on Wed May 12 21:26:30 2010
    Run ended on Wed May 12 21:27:06 2010
    Elapsed time was: 00:00:36.08
    CPU time was: 00:00:00.00

  • Problem: load SQL Server 'varbinary' data

    Source database: SQL Server 2000
    OS: Windows 2003 (SP1)
    Oracle: 10g (R2)
    Datatype Mapping: SQL Server ‘varbinary(1000)’ to oracle ‘BLOB’
    With the help of OMWB, I created the oracle database schema for a SQL Server DB (offline capture). I have problems when I tried to populate the (destination) database with a table with ‘varbinary(1000)’ datatype:
    PlateX.expID (as on http://www.sdss.org.uk/mybestdr5/en/help/browser/description.asp?n=PlateX&t=U) and the with original data is like this: http://dsg.port.ac.uk/~hx/research/sdss/logs/SelectExpidFromPlatex
    The PLATEX.ctl file (generated by OMWB) is like this (http://dsg.port.ac.uk/~hx/research/sdss/logs/PLATEX.ctl):
    load data
    infile 'PLATEX.dat' "str '<er>'"
    into table PLATEX
    fields terminated by '<ec>'
    trailing nullcols
    <SKIP>
    EXPID)
    I failed to load the data with the sql_load_script.sh script, and the log file is like this:
    <SKIP>
    EXPID NEXT * CHARACTER
    Terminator string : '<ec>'
    value used for ROWS parameter changed from 64 to 12
    Record 1: Rejected - Error on table PLATEX, column EXPID.
    Field in data file exceeds maximum length
    <SKIP>
    Record 60: Rejected - Error on table PLATEX, column EXPID.
    Field in data file exceeds maximum length
    MAXIMUM ERROR COUNT EXCEEDED - Above statistics reflect partial run.
    Table PLATEX:
    9 Rows successfully loaded.
    51 Rows not loaded due to data errors.
    0 Rows not loaded because all WHEN clauses were failed.
    0 Rows not loaded because all fields were null.
    Space allocated for bind array: 238392 bytes(12 rows)
    Read buffer bytes: 1048576
    Total logical records skipped: 0
    Total logical records read: 60
    Total logical records rejected: 51
    Total logical records discarded: 0
    Any help would be extremely appreciated.
    Helen
    P.S. previously, I mapped: SQL Server ‘varbinary(1000)’ to ‘RAW(1000)’ in Oracle but failed to populate the data (http://dsg.port.ac.uk/~hx/research/sdss/logs/PLATEX.log).

    Hi Emile,
    regading extracting data from MSSQL with OWB on Unix platform (using Generic Connectivity):
    Re: SQLServer access from AIX Warehouse builder
    Re: OWB on Solaris Connectivity with SQL SERVER on Windows
    We want to load aprox. 100 million records a day.
    I've read some articles about this and the advise was to dump the data from SQL Server to files and load the files with OWB.100 million records per day is not a problem for daily extracting from MSSQL Server if you have 1-2 hour.
    In my opinion dumping to text file is a bad practice and is unnecessary if customer don't have special requirements (for example for security reason).
    SQL Server source table and a Oracle target table without any difficult transformationsIn my opninion the best way process data from MSSQL is to extract data to staging area (schema) on Oracle DB with mappings as simple as possible (ONLY filters, without any join), and most of data processing prefom in Staging area or during moving from staging to DWH.
    Also look at OWB user guide (how to use Generic Connectivity in OWB)
    http://download.oracle.com/docs/cd/E11882_01/owb.112/e10582/loading_ms_data.htm#i1064950
    Regards,
    Oleg
    Edited by: added link to OWB doc with description of using Generic Connectivity

  • Row is loaded (SQL*LOAD) while it should be rejected !

    Hello,
    I have an input data file that contains a date :12-03-2007
    I use a control file with SQL*LOADER that mentions :
    DATE_INPUTFILE DATE 'YYYY-MM-DD HH24:MI:SS' terminated by ";",
    I would expect the row to be rejected.
    But for some reason, it is getting loaded properly onto the database with the value : 20-Mar-0012 07:00:00 AM !
    Can someone tell me how to avoid this row to be loaded ?
    Thanks.

    Here is CTL file :
    LOAD DATA
    REPLACE
    INTO TABLE SMK_ITEM_TMP TRAILING NULLCOLS
    ID_ITEM terminated by ";",
    MODALITY terminated by ";",
    ITEM terminated by ";",
    ITEM_DESCRIPTION terminated by ";",
    SUPPLIER_NAME terminated by ";",
    MLS terminated by ";",
    ORIGIN terminated by ";",
    MINMAX terminated by ";",
    DESTINATION terminated by ";",
    SUPERMARKET terminated by ";",
    SSOID terminated by ";",
    DATE_MODIFIED DATE 'YYYY-MM-DD HH24:MI:SS' terminated by ";",
    SERIAL terminated by ";",
    LT terminated by ";",
    WHERE_USED terminated by ";",
    MODE_APPRO terminated by ";",
    LOT terminated by ";",
    NB_CARD terminated by ";",
    ON_HAND terminated by ";",
    CONSO terminated by ";",
    TREMBLAY terminated by ";",
    POIDS terminated by ";",
    LENGTH terminated by ";",
    LARGE terminated by ";",
    HEIGHT terminated by ";",
    QTY_PALETTE terminated by ";",
    QTY_UNIT terminated by ";",
    ICV terminated by ";",
    VOLTA terminated by ";",
    VELIZY terminated by ";",
    CONTROL_MORN terminated by ";",
    MIN_ORDER terminated by ";",
    FIX_QTY terminated by ";",
    MIN_QTY terminated by ";",
    MAX_QTY terminated by ";",
    DELIVERY terminated by ";",
    PALETTE terminated by ";",
    KITPK terminated by ";",
    ACTION terminated by ";"
    Here is result log file :
    SQL*Loader: Release 10.2.0.3.0 - Production on Wed Mar 14 12:25:02 2007
    Copyright (c) 1982, 2005, Oracle. All rights reserved.
    Control File: /opt/appl/smk/zzdevl/parm/otrdbt01/ctl/SMK_ITEM_TMP.CTL
    Data File: /opt/appl/smk/zzdevl/data/download/otrdbt01/SMK_ITEM.csv
    Bad File: /opt/appl/smk/zzdevl/logs/otrdbt01/SMK_ITEM_TMP.BAD
    Discard File: none specified
    (Allow all discards)
    Number to load: ALL
    Number to skip: 0
    Errors allowed: 999999
    Bind array: 512 rows, maximum of 65536 bytes
    Continuation: none specified
    Path used: Conventional
    Silent options: FEEDBACK
    Table "SMKADMIN"."SMK_ITEM_TMP", loaded from every logical record.
    Insert option in effect for this table: REPLACE
    TRAILING NULLCOLS option in effect
    Column Name Position Len Term Encl Datatype
    ID_ITEM FIRST * ; CHARACTER
    MODALITY NEXT * ; CHARACTER
    ITEM NEXT * ; CHARACTER
    ITEM_DESCRIPTION NEXT * ; CHARACTER
    SUPPLIER_NAME NEXT * ; CHARACTER
    MLS NEXT * ; CHARACTER
    ORIGIN NEXT * ; CHARACTER
    MINMAX NEXT * ; CHARACTER
    DESTINATION NEXT * ; CHARACTER
    SUPERMARKET NEXT * ; CHARACTER
    SSOID NEXT * ; CHARACTER
    DATE_MODIFIED NEXT * ; DATE YYYY-MM-DD HH24:MI:SS
    SERIAL NEXT * ; CHARACTER
    LT NEXT * ; CHARACTER
    WHERE_USED NEXT * ; CHARACTER
    MODE_APPRO NEXT * ; CHARACTER
    LOT NEXT * ; CHARACTER
    NB_CARD NEXT * ; CHARACTER
    ON_HAND NEXT * ; CHARACTER
    CONSO NEXT * ; CHARACTER
    TREMBLAY NEXT * ; CHARACTER
    POIDS NEXT * ; CHARACTER
    LENGTH NEXT * ; CHARACTER
    LARGE NEXT * ; CHARACTER
    HEIGHT NEXT * ; CHARACTER
    QTY_PALETTE NEXT * ; CHARACTER
    QTY_UNIT NEXT * ; CHARACTER
    ICV NEXT * ; CHARACTER
    VOLTA NEXT * ; CHARACTER
    VELIZY NEXT * ; CHARACTER
    CONTROL_MORN NEXT * ; CHARACTER
    MIN_ORDER NEXT * ; CHARACTER
    FIX_QTY NEXT * ; CHARACTER
    MIN_QTY NEXT * ; CHARACTER
    MAX_QTY NEXT * ; CHARACTER
    DELIVERY NEXT * ; CHARACTER
    PALETTE NEXT * ; CHARACTER
    KITPK NEXT * ; CHARACTER
    ACTION NEXT * ; CHARACTER
    value used for ROWS parameter changed from 512 to 6
    Table "SMKADMIN"."SMK_ITEM_TMP":
    1 Row successfully loaded.
    0 Rows not loaded due to data errors.
    0 Rows not loaded because all WHEN clauses were failed.
    0 Rows not loaded because all fields were null.
    Space allocated for bind array: 60372 bytes(6 rows)
    Read buffer bytes: 1048576
    Total logical records skipped: 0
    Total logical records read: 1
    Total logical records rejected: 0
    Total logical records discarded: 0
    Run began on Wed Mar 14 12:25:02 2007
    Run ended on Wed Mar 14 12:25:02 2007
    Elapsed time was: 00:00:00.88
    CPU time was: 00:00:00.04
    Here is the input file (cut & paste from an Excel)
    19381;BXV;2100405;Wrong Uppdate testPIGNON MOTEUR AT5 25 DENTS;BINDER MAGNETIC;frederic.xxnneson;1.672;t;RIP;t;100023935;12-03-2007;f;5;K7042-01;Min-Max;1;0;9;396;f;0.07;0;0;0;0;16;9;f;f;f;23;1;16;0;;f;t;U
    Thanks

  • SQL Loader: ignore data field

    Hi,
    in a SQL Loader control file, how do you specify that you do not wish to load a particular field?
    E.g. if I wanted to ignore the second field in the data file and load the first and third fields into a table, how would I modify the control file below?
    Thanks.
    load data
    infile 'example.dat'
    into table example
    fields terminated by ',' optionally enclosed by '"'
    col1 char,
    ignore char,
    col2 char
    )

    User FILLER keyword:
    col1 char,
    ignore filler char,
    col2 char
    )

  • SQL Loader, CLOB, delimited fields

    Hello.
    I have to load using SQL Loader data from csv file into table, which one field is CLOB type.
    Here is how ctl file initially looked like:
    UNRECOVERABLE
    LOAD DATA
    INFILE '.\csv_files\TSH_DGRA.csv'
    BADFILE '.\bad_files\TSH_DGRA.bad'
    DISCARDFILE '.\dsc_files\TSH_DGRA.dsc'
    APPEND
    INTO TABLE TSH_DGRA
    FIELDS TERMINATED BY ',' OPTIONALLY ENCLOSED BY '"'
    TRAILING NULLCOLS
    ID_OBJ_TSHD,
    PR_ZOOM_TSHD,
    PR_GRID_TSHD,
    PR_ELMGR_TSHD CHAR(4000) OPTIONALLY ENCLOSED BY '<clob>' AND '</clob>',
    PR_ALRMGR_TSHD CHAR(4000) OPTIONALLY ENCLOSED BY '<clob>' AND '</clob>'
    Problems are fields PR_ELMGR_TSHD and PR_ALRMGR_TSHD (CLOBs in table TSH_DGRA). Until data which should be loaded into CLOB fields are under 4000 characters long, it works fine, but what should I do if I want to load data which are longer than 4000 characters?
    If found on Link:[http://download.oracle.com/docs/cd/B14117_01/server.101/b10825/ldr_loading.htm#i1006803] which one sentence said that:
    "SQL*Loader defaults to 255 bytes when moving CLOB data, but a value of up to 2 gigabytes can be specified. For a delimited field, if a length is specified, that length is used as a maximum. If no maximum is specified, it defaults to 255 bytes. For a CHAR field that is delimited and is also greater than 255 bytes, you must specify a maximum length. See CHAR for more information about the CHAR datatype."
    So, my question is, how to specify "up to 2gb" as text said? I can not use CHAR datatype because it is limited to 4000 characters. And I have to load about 60000 characters. I also can not use technique where all data for every CLOB field are in separate files.

    Just specify the maximum expected size:
    PR_ELMGR_TSHD CHAR(100000) OPTIONALLY ENCLOSED BY '<clob>' AND '</clob>',
    PR_ALRMGR_TSHD CHAR(1000000) OPTIONALLY ENCLOSED BY '<clob>' AND '</clob>'
    The CHAR(1000000) will allow SQLLDR to handle up to 1000000 bytes of input text.

  • NOT ABLE TO SEE SQL EXPRESSION FIELDS IN CR2008

    Hi,
    I am not able to see the SQL EXPRESSION FIELDS in the FIELD EXPLORER of CR 2008.Can
    you please help me
    BABASHARI

    Hi,
    I would suggest you open the question in the Crystal Reports forum.
    Ingo

  • SQL Expression Fields button not visible in Field Explorer

    Hi,
    On my Crystal Reports installation I can see the button SQL Expression Fields in Field Explorer but it is not visible on client's installation.
    How do I make it visible? I need to create a new report with SQL query for client.
    Thanks,
    Ajay Audich

    Commands are intended to replace the standard linked tables that must users are accustomed to.
    I suppose Commands would be considered an "advanced" feature, considering that they require the developer to be able to hand write SQL (our at least copy it from another SQL editor).
    If you are comfortable writing reading, writing & editing SQL, Commands will open a the door to a whole new level of control when it comes to creating your data sets.
    The biggest problem w Commands is also it's greatest strength... If you're using a Command, CR will assume that you know what you're doing and will not try to rewrite or optimize your SQL in the background. You write it, CR sends it. That means if you write crappy SQL, CR will send your crappy SQL to the db server...
    It also means that you can use join types that aren't available in CR and use sub-queries in in your SQL... Basically anything you can do in the db's native editor.
    HTH,
    Jason

  • SQL Server 2012 Express - Could not load the file or assembly 'Microsoft.VisualStudio.OLE.Interop

    Windows 7 Home Premium, SP1
    64-bit
    I just installed SQL Server 2012 express - Express with Advanced Services (contains the database engine, Express Tools, Reporting Services, and Full Text Search),
    file name -  SQLEXPRADV_x64_ENU.
    Opening SQL Server Management Studio causes the error -
    "Could not load the file or assembly 'Microsoft.VisualStudio.OLE.Interop,
    Version=7.1.40304.0, Culture=neutral,
    PublicKeyToken=b03f5f7f11d50a3a' or one of its dependencies. The
    system cannot find the file specified."
    Do I need to install another file? Did I download the wrong version?
    Any suggestions? Thanks for any help.

    Hello,
    The issue may occure when the installation of your SQL Server is not a complete or a successful one. Did you try to repair or reinstall the SQL Server Express?  You can also try to install
    Service Pack 1 for SQL Server 2012
    to check if the issue presists.
    Regards,
    Fanny Liu
    Fanny Liu
    TechNet Community Support

  • Main Table data load u2013 UNSPSC fields is not loading

    I am new to SAP MDM
    I have the main table data that includes UNSPSC field. UNSPSC (hierarchy) table is already loaded.
    It works fine when I use import manager with field mapping and value mapping. (UNSPSC field value mapping is done).
    When I use the import server using the same map to load the main table data with UNSPSC field (in this case the UNSPSC field value is different but UNSPSC lookup table has that value) , UNSPSC field is not loading but all other fields are loaded including images and PDF's with new values
    If I go to the import manager and do the value mapping again for the UNSPSC field with the new value then save the map and use the import server to load the data then it is loading correctly.
    My question when we use the import server, main table data with UNSPSC codes value will be different  each time and it doesnu2019t make sense to go to the import manager and do the value mapping and saving the import map  before loading the data again.
    What I am missing here?.  Anyone can help me?

    Could anyone clarify this?
    Issue: UNSPSC field value mapping automatically by using the import server while loading the Main table.
    This issue was resolved yesterday and still works fine with the remote system MDC UNSPSC.
    Is there anyn settings in the ' Set MDIS Unmapped value handling'? (Right click on the field Product hierarchy  field at the destination side). By default it is setting to 'Add' for both the working remote system as well as the non working remote system
    SAP MDM 5.5 SP6 and I am using the standard Product Master repository
    I tried this in a different remote system MDC R/3 & ERP and it worked some time and didnu2019t work later. If it is working then during the UNSPSC code field mapping,  it automatically maps the values also.
    The destination side the main table Products and the destination side [Remote key] field is displayed.
    Source file, I have only 3 fields and they are Product No, Product Name and UNSPSC Category and UNSPSC Category is mapped to the destination field Product Hierarchy field(lookup hierarchy)
    Do I have to map any field  or clone any field and map to the [Remote Key Field]  in the destination side? If yes, what field I have to clone and map it to the Remote Key filed? Is there any other settings necessary. I am not using any matching with this field or any other field.
    Steve.
    Edited by: SteveLat on Oct 8, 2009 11:57 PM
    Edited by: SteveLat on Oct 9, 2009 12:03 AM
    Edited by: SteveLat on Oct 9, 2009 12:47 AM

  • System.Data.SqlServerCe.SqlCeException (0x80004005): Native components of SQL Server Compact engine are not loaded. Please reinstall Microsoft SQL Server Compact.

    Hi
    I have followed http://technet.microsoft.com/en-us/library/ms173298(v=sql.110).aspx to support private deployment for my application in anycpu mode.
    Though its working fine on several XP and Windows 7 , on on particular XP machine , I get this
    System.Data.SqlServerCe.SqlCeException (0x80004005): Native components of SQL Server Compact engine are not loaded. Please reinstall Microsoft SQL Server Compact.
       at System.Data.SqlServerCe.NativeMethods.ThrowIfNativeLibraryNotLoaded()
       at System.Data.SqlServerCe.SqlCeEngine.Dispose(Boolean disposing)
       at System.Data.SqlServerCe.SqlCeEngine.Finalize()
    Can anybody give me directions on this?

    Do you use LINQ to SQL or EF or ADO.NET, is what 
    Yes, I would start by investigating:
    - Do you REALLY need "any CPU", or can you cope with just x86 to make things simpler (See my LINQ to SQL blog post)
    - Is SQL Server Compact installed on this PC, and which build (is it same or different from the one you distribute)
    - Use Process Explorer from sysinternals on the machine to find out where the app tries to load the native libraries from (and fails)
    Please mark as answer, if this was it. Visit my SQL Server Compact blog http://erikej.blogspot.com

  • SQL Expression fields do not display when no records are returned

    I have a report that uses SQl expression fields to summarize in the Report footer.  When the main report does not contain any records the SQL expression fields are blank.
    Is there a way to force these SQL expression fields to be evaluated evn when the main report does not contain any records.
    Thanks

    Hi,
    You can try this :
    Under Report Options, check "Convert Database Null values to Default".

Maybe you are looking for

  • USER STATUS at Activity Level

    I do have user status defined for Activity and it doesn't show up when i set the status. All i see is SYSTEM STATUS and not user status. To see user status i need to click the blue button (with I symbol) and then see it. Why is this happening like th

  • Web Application services failing to start (Foundation SS, Hyperion Admin)

    Hello, Hyperion Essbase 11.1.1.3 In my installation recently my web applications are failing to start properly. I'm following the start order and allowing services time to come up, but they are still reporting an error. 1. OpenLDAP starts fine, and q

  • What's the maximum number of Apple TVs that can be displayed in the AirPlay list?

    We have 50 conference rooms all on the same network and are considering equiping each one with an Apple TV for AirPlay Mirroring our iPad 2s to the rooms projector, or TV. What is the maximum number of Apple TVs that can be displayed in the AirPlay m

  • Few Questions about macbook pros

    Is the transition from windows to mac easy or does it take awhile? Did you previous windows users get used to the minimizing of windows,I find it alittle weird? Another question,Why do people ask if they should get a macbook or macbook pro if price a

  • Operation timed out (FSP network services manager)

    Received this error from five different RMAN jobs, all submitted at approx 6am, all failing within a couple of minutes. Resubmit of all five jobs worked fine. Many other jobs submitted at approx the same time worked fine. Time out does not appear to