How to avoid the split problem when uploading the data from csv file

Dear Friends,
              I have to upload data from the .csv file to my custom table , i have found a problem when uploading the data .
i am using the code as below , please suggest me what i have to do in this regard
      SPLIT wa_raw_csv  AT ',' INTO
                wa_empdata_csv-status
                 wa_empdata_csv-userid
                 wa_empdata_csv-username
                 wa_empdata_csv-Title
                 wa_empdata_csv-department.
APPEND wa_empdata_csv TO  itab.
in the flat file i can see for one of the record for the field Title  as
Director, Finance - NAR............there by through my code the  wa_empdata_csv-Title is getting splited data as "Director, and  Department field is getting Finance - NAR" , i can see that even though  " Director, Finance - NAR"  is one word it is getting split .
.......which is the problem iam facing.Please could any body let me know how in this case i should handle in my code that this word
"Director,Finance - NAR"   wil not be split into two words.
Thanks & Records
Madhuri

Hi Madhuri,
Best way to avoid such problem is to use TAB delimited file instead of comma separated data. Generally TAB does not appear in data.
If you are generating the file, use tab instead of comma.
If you cannot modify the format of file and data length in file is fixed character, you will need to define the structure and then move data in fixed length structure.
Regards,
Mohaiyuddin

Similar Messages

  • How to avoid invoice split due to different billing dates?

    Hi All,
    How to avoid invoice split due to different billing dates during executing VF04 transction?
    I want to club 30 delivery documents with different billing due dates in one invoice.
    I am using VF04 tr for the same. However, here i cannot control the invoice split based on diff billing due dates.
    If the same thing is tried via VF01 , it allows only 14 del dcouments at a time.
    Can u suggest  abetter technique to do the same?
    Regards,
    Niketan

    I think I already explained you in a similar post to use VF04 setting the billing date as default:
    menu: settings->default data. Place  the date in field Billing Date.
    pls. reward if helpfull
    Roberto

  • Can I specify the bank account when uploading the bank statement?

    Currently when I use FF.5 to upload the bank statement, SAP deteremine the the bank G/L account to do posting accoring to the bank and account information in the bank statement. Do we have other choice? Like, no matter with the bank information in the bank statement, we can specify the account manually when uploading the bank statement?

    Hi,
    No you cannot specify the G/L accounts manually while uploading the bank statement. They are derived from the config done in Fin Acc > Bank Acc > Bus Tr > Payt tr > EBS > Make global sett....
    Br, NK

  • How to load the data from .csv file to oracle table???

    Hi,
    I am using oracle 10g , plsql developer. Can anyone help me in how to load the data from .csv file to oracle table. The table is already created with the required columns. The .csv file is having about 10lakh records. Is it possible to load 10lakh records. can any one please tell me how to proceed.
    Thanks in advance

    981145 wrote:
    Can you tell more about sql * loader??? how to know that utility is available for me or not??? I am using oracle 10g database and plsql developer???SQL*Loader is part of the Oracle client. If you have a developer installation you should normally have it on your client.
    the command is
    sqlldrType it and see if you have it installed.
    Have a look also at the FAQ link posted by Marwin.
    There are plenty of examples also on the web.
    Regards.
    Al

  • How to create  a datasource for 0COSTCENTER to load data in csv file  in BI

    how to create  a datasource for 0COSTCENTER to load data in csv file  in BI 7.0 system
    can you emil me the picture of the step about how to  loaded individual values of the hierarchy using CSV file
    thank you very much
    my emil is <Removed>
    allen

    Step 1: Load Required Master Data for 0CostCenter in to BI system
    Step 2: Enable Characteristics to support Hierarchy for this 0Cost Center and specify the External Characteristic(In the Lowest Node or Last Node) while creation of this Characteristic InfoObject
    Step 3: On Last Node of Hierarchy Structure in the InfoObject, Right Click and then Create Hierarchy MANUALLY by Inserting the Master Data Value as BI dosent Support the Hierarchy load directly you need to do it manually....
    Step 4: Mapping
    Create Text Node thats the first node (Root Node)
    Insert Characteristic Nodes
    Insert the Last Node of the Hierarchy
    Then you need to create a Open hub Destination for extracting data into the .csv file...
    Step1 : Create the Open Hub Destination give the Master Data table name and enter all the fields required....and create the transformations for this Open Hub connecting to the External file or excel file source...then give the location on to your local disk or path of the server in the first tab and request for the data...It should work alright let me know if you need anything else...
    Thanks,
    Sandhya

  • Error when executing interface which load data from csv file which has 320

    Hi,
    Can some one provide a resolution for below error:
    I have created an interface which load data from csv file which has 320 columns, to a Synonym which has 320 columns in it
    using LKM File to SQL, IKM Sql Control Append.
    I am getting below error when executing the interface :
    com.sunopsis.tools.core.exception.SnpsSimpleMessageException: ODI-17517: Error during task interpretation. Task: 6 java.lang.Exception: BeanShell script error: Sourced file: inline evaluation of: ``out.print("The application script threw an exception: java.lang.StringIndexOutOf . . . '' Token Parsing Error: Lexical error at line 2, column 42. Encountered: "\\" (92), after : "": <at unknown location> BSF info: Create external table at line: 0 column: columnNo
    at com.sunopsis.dwg.codeinterpretor.SnpCodeInterpretor.transform(SnpCodeInterpretor.java:485)
         at com.sunopsis.dwg.dbobj.SnpSessStep.createTaskLogs(SnpSessStep.java:711)
         at com.sunopsis.dwg.dbobj.SnpSessStep.treatSessStep(SnpSessStep.java:461)
         at com.sunopsis.dwg.dbobj.SnpSession.treatSession(SnpSession.java:2093)
         at oracle.odi.runtime.agent.processor.impl.StartSessRequestProcessor$2.doAction(StartSessRequestProcessor.java:366)
         at oracle.odi.core.persistence.dwgobject.DwgObjectTemplate.execute(DwgObjectTemplate.java:216)
         at oracle.odi.runtime.agent.processor.impl.StartSessRequestProcessor.doProcessStartSessTask(StartSessRequestProcessor.java:300)
         at oracle.odi.runtime.agent.processor.impl.StartSessRequestProcessor.access$0(StartSessRequestProcessor.java:292)
         at oracle.odi.runtime.agent.processor.impl.StartSessRequestProcessor$StartSessTask.doExecute(StartSessRequestProcessor.java:855)
         at oracle.odi.runtime.agent.processor.task.AgentTask.execute(AgentTask.java:126)
         at oracle.odi.runtime.agent.support.DefaultAgentTaskExecutor$2.run(DefaultAgentTaskExecutor.java:82)
         at java.lang.Thread.run(Thread.java:662)
    Caused by: org.apache.bsf.BSFException: BeanShell script error: Sourced file: inline evaluation of: ``out.print("The application script threw an exception: java.lang.StringIndexOutOf . . . '' Token Parsing Error: Lexical error at line 2, column 42. Encountered: "\\" (92), after : "": <at unknown location>
    BSF info: Create external table at line: 0 column: columnNo
         at bsh.util.BeanShellBSFEngine.eval(Unknown Source)
         at bsh.util.BeanShellBSFEngine.exec(Unknown Source)
         at com.sunopsis.dwg.codeinterpretor.SnpCodeInterpretor.transform(SnpCodeInterpretor.java:471)
         ... 11 more
    Text: The application script threw an exception: java.lang.StringIndexOutOfBoundsException: String index out of range: 2 BSF info: Create external table at line: 0 column: columnNo
    out.print("createTblCmd = r\"\"\"\ncreate table ") ;
    out.print(odiRef.getTable("L", "COLL_NAME", "W")) ;
    out.print("<?=(extTabColFormat.getUseView())?\"_ET\":\"\"?>\n(\n\t") ;
    out.print(odiRef.getColList("", "[CX_COL_NAME]\\t"+
              "<?=extTabColFormat.getExtTabDataType(\\u0022[CX_COL_NAME]\\u0022,\\u0022[SOURCE_DT]\\u0022, \\u0022[DEST_WRI_DT]\\u0022, \\u0022[COL_FORMAT]\\u0022, \\u0022[BYTES]\\u0022, \\u0022[LONGC]\\u0022, \\u0022[SCALE]\\u0022)?>"
         , ",\\n\\t", "","")) ;
    out.print("\n)\nORGANIZATION EXTERNAL\n(\n\tTYPE ORACLE_LOADER\n\tDEFAULT DIRECTORY dat_dir\n\tACCESS PARAMETERS\n\t(\n\t\tRECORDS DELIMITED BY 0x'") ;
    out.print(odiRef.getSrcTablesList("[XFILE_SEP_ROW]","")) ;
    out.print("'\n\t\t") ;
    out.print(odiRef.getUserExit("EXT_CHARACTERSET")) ;
    out.print("\n\t\t") ;
    out.print(odiRef.getUserExit("EXT_STRING_SIZE")) ;
    out.print("\n\t\tBADFILE\t\t'") ;
    out.print(odiRef.getSrcTablesList("", "[RES_NAME]", "", "")) ;
    out.print("_%a.bad'\n\t\tLOGFILE\t\t'") ;
    out.print(odiRef.getSrcTablesList("", "[RES_NAME]", "", "")) ;
    out.print("_%a.log'\n\t\tDISCARDFILE\t'") ;
    out.print(odiRef.getSrcTablesList("", "[RES_NAME]", "", "")) ;
    out.print("_%a.dsc'\n\t\tSKIP \t\t") ;
    out.print(odiRef.getSrcTablesList("", "[FILE_FIRST_ROW]", "", "")) ;
    out.print("\n") ;
    if (odiRef.getSrcTablesList("", "[FILE_FORMAT]", "", "").equals("F")) {out.print("\n\t\tFIELDS\n\t\t") ;
    out.print(odiRef.getUserExit("EXT_MISSING_FIELD")) ;
    out.print("\n\t\t(\n\t\t\t") ;
    out.print(odiRef.getColList("", "[CX_COL_NAME]\\tPOSITION([FILE_POS]:[FILE_END_POS])\\t"+
                        "<?=extTabColFormat.getExtTabFormat(\\u0022[CX_COL_NAME]\\u0022,\\u0022[SOURCE_DT]\\u0022, \\u0022DEST_WRI_DT\\u0022, \\u0022[COL_FORMAT]\\u0022, \\u0022[BYTES]\\u0022, \\u0022[LONGC]\\u0022, \\u0022[SCALE]\\u0022)?>"
                        , ",\\n\\t\\t\\t", "","")) ;
    out.print("\t\t\n\t\t)\n\t)\n") ;
    } else {out.print("\n\t\tFIELDS TERMINATED BY x'") ;
    out.print(odiRef.getSrcTablesList("", "[XFILE_SEP_FIELD]", "", "")) ;
    out.print("'\n\t\t") ;
    if(odiRef.getSrcTablesList("", "[FILE_ENC_FIELD]", "", "").equals("")){out.print("\n\t\t") ;
    } else {out.print("OPTIONALLY ENCLOSED BY '") ;
    out.print(odiRef.getSrcTablesList("", "[FILE_ENC_FIELD]", "", "").substring(0,1)) ;
    out.print("' AND '") ;
    out.print(odiRef.getSrcTablesList("", "[FILE_ENC_FIELD]", "", "").substring(1,2)) ;
    out.print("' ") ;
    }out.print("\n\t\t") ;
    out.print(odiRef.getUserExit("EXT_MISSING_FIELD")) ;
    out.print("\n\t\t(\n\t\t\t") ;
    out.print(odiRef.getColList("", "[CX_COL_NAME]\\t"+
                        "<?=extTabColFormat.getExtTabFormat(\\u0022[CX_COL_NAME]\\u0022,\\u0022[SOURCE_DT]\\u0022, \\u0022DEST_WRI_DT\\u0022, \\u0022[COL_FORMAT]\\u0022, \\u0022[BYTES]\\u0022, \\u0022[LONGC]\\u0022, \\u0022[SCALE]\\u0022)?>"
                        , ",\\n\\t\\t\\t", "","")) ;
    out.print("\t\t\n\t\t)\n\t)\n") ;
    }out.print("\tLOCATION (") ;
    out.print(odiRef.getSrcTablesList("", "'[RES_NAME]'", "", "")) ;
    out.print(")\n)\n") ;
    out.print(odiRef.getUserExit("EXT_PARALLEL")) ;
    out.print("\nREJECT LIMIT ") ;
    out.print(odiRef.getUserExit("EXT_REJECT_LIMIT")) ;
    out.print("\n\"\"\"\n \n# Create the statement\nmyStmt = myCon.createStatement()\n \n# Execute the trigger creation\nmyStmt.execute(createTblCmd)\n \nmyStmt.close()\nmyStmt = None\n \n# Commit, just in case\nmyCon.commit()") ;
    ****** ORIGINAL TEXT ******
    createTblCmd = r"""
    create table <%=odiRef.getTable("L", "COLL_NAME", "W")%><?=(extTabColFormat.getUseView())?"_ET":""?>
         <%=odiRef.getColList("", "[CX_COL_NAME]\t"+
              "<?=extTabColFormat.getExtTabDataType(\u0022[CX_COL_NAME]\u0022,\u0022[SOURCE_DT]\u0022, \u0022[DEST_WRI_DT]\u0022, \u0022[COL_FORMAT]\u0022, \u0022[BYTES]\u0022, \u0022[LONGC]\u0022, \u0022[SCALE]\u0022)?>"
         , ",\n\t", "","")%>
    ORGANIZATION EXTERNAL
         TYPE ORACLE_LOADER
         DEFAULT DIRECTORY dat_dir
         ACCESS PARAMETERS
              RECORDS DELIMITED BY 0x'<%=odiRef.getSrcTablesList("[XFILE_SEP_ROW]","")%>'
              <%=odiRef.getUserExit("EXT_CHARACTERSET")%>
              <%=odiRef.getUserExit("EXT_STRING_SIZE")%>
              BADFILE          '<%=odiRef.getSrcTablesList("", "[RES_NAME]", "", "")%>_%a.bad'
              LOGFILE          '<%=odiRef.getSrcTablesList("", "[RES_NAME]", "", "")%>_%a.log'
              DISCARDFILE     '<%=odiRef.getSrcTablesList("", "[RES_NAME]", "", "")%>_%a.dsc'
              SKIP           <%=odiRef.getSrcTablesList("", "[FILE_FIRST_ROW]", "", "")%>
    <% if (odiRef.getSrcTablesList("", "[FILE_FORMAT]", "", "").equals("F")) {%>
              FIELDS
              <%=odiRef.getUserExit("EXT_MISSING_FIELD")%>
                   <%=odiRef.getColList("", "[CX_COL_NAME]\tPOSITION([FILE_POS]:[FILE_END_POS])\t"+
                        "<?=extTabColFormat.getExtTabFormat(\u0022[CX_COL_NAME]\u0022,\u0022[SOURCE_DT]\u0022, \u0022DEST_WRI_DT\u0022, \u0022[COL_FORMAT]\u0022, \u0022[BYTES]\u0022, \u0022[LONGC]\u0022, \u0022[SCALE]\u0022)?>"
                        , ",\n\t\t\t", "","")%>          
    <%} else {%>
              FIELDS TERMINATED BY x'<%=odiRef.getSrcTablesList("", "[XFILE_SEP_FIELD]", "", "")%>'
              <% if(odiRef.getSrcTablesList("", "[FILE_ENC_FIELD]", "", "").equals("")){%>
              <%} else {%>OPTIONALLY ENCLOSED BY '<%=odiRef.getSrcTablesList("", "[FILE_ENC_FIELD]", "", "").substring(0,1)%>' AND '<%=odiRef.getSrcTablesList("", "[FILE_ENC_FIELD]", "", "").substring(1,2)%>' <%}%>
              <%=odiRef.getUserExit("EXT_MISSING_FIELD")%>
                   <%=odiRef.getColList("", "[CX_COL_NAME]\t"+
                        "<?=extTabColFormat.getExtTabFormat(\u0022[CX_COL_NAME]\u0022,\u0022[SOURCE_DT]\u0022, \u0022DEST_WRI_DT\u0022, \u0022[COL_FORMAT]\u0022, \u0022[BYTES]\u0022, \u0022[LONGC]\u0022, \u0022[SCALE]\u0022)?>"
                        , ",\n\t\t\t", "","")%>          
    <%}%>     LOCATION (<%=odiRef.getSrcTablesList("", "'[RES_NAME]'", "", "")%>)
    <%=odiRef.getUserExit("EXT_PARALLEL")%>
    REJECT LIMIT <%=odiRef.getUserExit("EXT_REJECT_LIMIT")%>
    # Create the statement
    myStmt = myCon.createStatement()
    # Execute the trigger creation
    myStmt.execute(createTblCmd)
    myStmt.close()
    myStmt = None
    # Commit, just in case
    myCon.commit().
         at com.sunopsis.dwg.dbobj.SnpSessStep.createTaskLogs(SnpSessStep.java:738)
         at com.sunopsis.dwg.dbobj.SnpSessStep.treatSessStep(SnpSessStep.java:461)
         at com.sunopsis.dwg.dbobj.SnpSession.treatSession(SnpSession.java:2093)
         at oracle.odi.runtime.agent.processor.impl.StartSessRequestProcessor$2.doAction(StartSessRequestProcessor.java:366)
         at oracle.odi.core.persistence.dwgobject.DwgObjectTemplate.execute(DwgObjectTemplate.java:216)
         at oracle.odi.runtime.agent.processor.impl.StartSessRequestProcessor.doProcessStartSessTask(StartSessRequestProcessor.java:300)
         at oracle.odi.runtime.agent.processor.impl.StartSessRequestProcessor.access$0(StartSessRequestProcessor.java:292)
         at oracle.odi.runtime.agent.processor.impl.StartSessRequestProcessor$StartSessTask.doExecute(StartSessRequestProcessor.java:855)
         at oracle.odi.runtime.agent.processor.task.AgentTask.execute(AgentTask.java:126)
         at oracle.odi.runtime.agent.support.DefaultAgentTaskExecutor$2.run(DefaultAgentTaskExecutor.java:82)
         at java.lang.Thread.run(Thread.java:662)

    The issue is encountered because the text delimiter used in the source file did not consist of a pair of delimiters.
    Please see support Note [ID 1469977.1] for details.

  • Help Me... How to get Data from CSV File...?

    Hi Everyone..!
    This is yajiv and am working in CS3 Photoshop platform. I know about Java Script. Is it possible to get the data from CSV files. Actually our client use to send us the CSV files which contains a lot of swatch name and reference files in one particular image name.
    Actually how we work on that CSV file is, first we copy file name to search that CSV file. then get the result to paste into layer name. This process continue till the end of swatch.
    Thank in Advance
    -yajiv

    > Is it possible to get the data from CSV files.
    Have you tried searching this forum?

  • Problem when upload legacy data

    Hi all,
    I've faced a problem when uploading asset legacy data (AS91) for customer. My customer have started depreciation on May 21, 2009, and I set transfer date is November 30, 2010. So SAP doesn't calculate right planned value for December, 2010. I think that problem related to the first month depreciation. It calculated full month for depreciation (instead of from May 21) and change the planned value for December 2010. So I want SAP just calculate 11 days depreciation for May 2009. How can I config to do this?
    Thanks in advance!

    Hi,
    Check in the asset master Opening Start Date for depreciation.. If it is under open mode change that date and see asset explorer for depreciation values.
    Thanks,
    Padmaja

  • Issue when uploading Sales data from DSO to Cube.

    Dear All,
    I have an issue when I am uploading Sales data from DSO to Cube.  I am using BI 7.0 and I have uploaded all sales document level data to my DSO.  Then I use a transformation rule to calculate the Sales Value when I am doing DTP to cube.  Cube has customer wise aggregation data.
    In DSO I have NetPrice(KF) and Delivered_QTY(KF).  I do a simple multiplication routine in the transformation from DSO to Cube.
    RESULT =   SOURCE_FIELDS-NET_PRICE * SOURCE_FIELDS-DLV_QTY .
    At the moment I use Active Table(With out Archive) on the DSO to get the data since this is my first load.
    Issue is that the figure (Sales Value) in the cube is incorrect. I am getting a very large values which is impossible.
    Can someone please help me.
    Shanka

    Hi,
    are you sure that cube has customer wise aggregated data..since it will always aggregate the values of the key figures for the same set of characters...
    did you checked the values for the key figures as well and if even they are inflated or is the problem with this key figure only.
    During the data load the records may aggregates first for the same values of characters and then multiplication will be happening....if that is the case then you may have to multiply the values before storing it in the data package and then let it aggregate....can be achived through start routine...
    but first verify if other key figures are also having the same issue.
    Thanks
    Ajeet

  • How to import data from CSV file with columns separated by semicolon?

    I migrate database from MS SQL 2008 to ORACLE 11g
    I export data to CSV file from MS SQL
    then I try to import it to Oracle
    several tables goes fine using Import data option in the SQL Developer
    Standard CSV files with data separated by comma were imported.
    chars, date (with format string), and integer data are imported via import wizard without problems
    the problems were when I try to import table with noninteger numbers with modal part separated by comma not by dot
    comma is the standard separator for columns in CSV file
    so I must change the standard separator to semicolon
    then the import wizard have problem to correct recognize the columns data because it use only standard CSV comma separator :-/
    In SQL Developer 1.5.3 Tools -> Preferences -> Migration -> Data Move Options
    I change "End of Column Delimiter" to ; but it doens't work
    Is this possible to change the standard column separator for import data wizzard in SQL Developer 1.5.3?
    Or maybe someone know how to import data in SQL Developer 1.5.3 from CSV when CSV colunn separator is set to semicolon?

    A new preference has been added to customize the import delimiter in main code line. This should be available as part of future release.

  • Problem in converting table data into CSV file

    Hi All,
    In my Process i need to convert my error table data into csv file,my data is converted as csv file by using OdisqlUnload function,but the column headers are not converted,i use another procedure for converting column headers but iam getting below error ...
    com.sunopsis.sql.SnpsMissingParametersException: Missing parameter string.find, string.find
    SQL: import string import java.sql as sql import java.lang as lang import re sourceConnection = odiRef.getJDBCConnection("SRC") output_write=open('C:/Oracle/Middleware/Oracle_ODI2/oracledi/pro/PRO.txt','r+') myStmt = sourceConnection.createStatement() my_query = "select * FROM E$_LOCAL_F0911Z1" my_query=my_query.upper() if string.find(my_query, '*') > 0: myRs = myStmt.executeQuery(my_query) md=myRs.getMetaData() collect=[] i=1 while (i <= md.getColumnCount()): collect.append(md.getColumnName(i)) i += 1 header=','.join(map(string.strip, collect)) elif string.find(my_query,'||') > 0: header = my_query[7:string.find(my_query, 'FROM')].replace("||','||",',') else: header = my_query[7:string.find(my_query, 'FROM')] print header old=output_write.read() output_write.seek(0) output_write.write (header+'\n'+old) sourceConnection.close() output_write.close()
    And i used below code for converting.......
    import string
    import java.sql as sql
    import java.lang as lang
    import re
    sourceConnection = odiRef.getJDBCConnection("SRC")
    output_write=open('C:/Oracle/Middleware/Oracle_ODI2/oracledi/pro/PRO.txt','r+')
    myStmt = sourceConnection.createStatement()
    my_query = "select FROM E$_COMPANY"*
    *my_query=my_query.upper()*
    *if string.find(my_query, '*') > 0:*
    *myRs = myStmt.executeQuery(my_query)*
    *md=myRs.getMetaData()*
    *collect=[]*
    *i=1*
    *while (i <= md.getColumnCount()):*
    *collect.append(md.getColumnName(i))*
    *i += 1*
    *header=','.join(map(string.strip, collect))*
    *elif string.find(my_query,'||') > 0:*
    *header = my_query[7:string.find(my_query, 'FROM')].replace("||','||",',')*
    *else:*
    *header = my_query[7:string.find(my_query, 'FROM')]*
    *print header*
    *old=output_write.read()*
    *output_write.seek(0)*
    *output_write.write (header+'\n'+old)*
    *sourceConnection.close()*
    *output_write.close()*
    Any one can you help regarding this
    Edited by: 30021986 on Oct 1, 2012 6:04 PM

    This may not be an option for you but in pinch you may want to consider outputing your data to an MS Spreadsheet, then saving it as a CSV. It's somewhat of a cumbersome process, but it will get you by for now.
    You will need to change your content type to application/vnd.ms-excel.
    <% response.setContentType("application/vnd.ms-excel"); %>

  • Problem when simultaneously retriving data from berkeleydb using web applic

    Hi all
    Plz Help
    i am using jboss server .
    my problem is :
    i have 2lacks recodrs in berkeleydatabase
    when i try to retrive data it working properly for one browser
    if i start retrive data simultaneously using two IES it getting this exception
    how to solve this problem,i mean separeate session for separate client
    in java session trking techique i applied ,but here not working retriving data from berkeleydb.collision occur
    11:47:08,468 ERROR [STDERR] Database handles still open at environment close
    11:47:08,468 ERROR [STDERR] Open database handle: smscsucc.db
    11:47:08,468 ERROR [STDERR] java.lang.IllegalArgumentException: Invalid argument
    11:47:08,468 ERROR [STDERR] at com.sleepycat.db.internal.db_javaJNI.DbEnv_cl
    ose0(Native Method)
    11:47:08,468 ERROR [STDERR] at com.sleepycat.db.internal.DbEnv.close0(DbEnv.
    java:208)
    11:47:08,468 ERROR [STDERR] at com.sleepycat.db.internal.DbEnv.close(DbEnv.j
    ava:76)
    11:47:08,468 ERROR [STDERR] at com.sleepycat.db.Environment.close(Environmen
    t.java:39)
    11:47:08,468 ERROR [STDERR] at daos.RetriveDao.getData(RetriveDao.java:131)
    11:47:08,468 ERROR [STDERR] at controllers.MainServlet.getData(MainServlet.j
    ava:47)
    11:47:08,468 ERROR [STDERR] at controllers.MainServlet.serviceHelper(MainSer
    vlet.java:76)
    11:47:08,468 ERROR [STDERR] at controllers.MainServlet.doGet(MainServlet.jav
    a:101)
    11:47:08,468 ERROR [STDERR] at javax.servlet.http.HttpServlet.service(HttpSe
    rvlet.java:697)
    11:47:08,468 ERROR [STDERR] at javax.servlet.http.HttpServlet.service(HttpSe
    rvlet.java:810)
    11:47:08,468 ERROR [STDERR] at org.apache.catalina.core.ApplicationFilterCha
    in.internalDoFilter(ApplicationFilterChain.java:252)
    11:47:08,468 ERROR [STDERR] at org.apache.catalina.core.ApplicationFilterCha
    in.doFilter(ApplicationFilterChain.java:173)
    11:47:08,468 ERROR [STDERR] at org.jboss.web.tomcat.filters.ReplyHeaderFilte
    r.doFilter(ReplyHeaderFilter.java:96)
    11:47:08,468 ERROR [STDERR] at org.apache.catalina.core.ApplicationFilterCha
    in.internalDoFilter(ApplicationFilterChain.java:202)
    11:47:08,468 ERROR [STDERR] at org.apache.catalina.core.ApplicationFilterCha
    in.doFilter(ApplicationFilterChain.java:173)
    11:47:08,468 ERROR [STDERR] at org.apache.catalina.core.StandardWrapperValve
    .invoke(StandardWrapperValve.java:213)
    11:47:08,468 ERROR [STDERR] at org.apache.catalina.core.StandardContextValve
    .invoke(StandardContextValve.java:178)
    11:47:08,468 ERROR [STDERR] at org.jboss.web.tomcat.security.SecurityAssocia
    tionValve.invoke(SecurityAssociationValve.java:175)
    11:47:08,468 ERROR [STDERR] at org.jboss.web.tomcat.security.JaccContextValv
    e.invoke(JaccContextValve.java:74)
    11:47:08,468 ERROR [STDERR] at org.apache.catalina.core.StandardHostValve.in
    voke(StandardHostValve.java:126)
    11:47:08,468 ERROR [STDERR] at org.apache.catalina.valves.ErrorReportValve.i
    nvoke(ErrorReportValve.java:105)
    11:47:08,468 ERROR [STDERR] at org.apache.catalina.core.StandardEngineValve.
    invoke(StandardEngineValve.java:107)
    11:47:08,468 ERROR [STDERR] at org.apache.catalina.connector.CoyoteAdapter.s
    ervice(CoyoteAdapter.java:148)
    11:47:08,468 ERROR [STDERR] at org.apache.coyote.http11.Http11Processor.proc
    ess(Http11Processor.java:869)
    11:47:08,468 ERROR [STDERR] at org.apache.coyote.http11.Http11BaseProtocol$H
    ttp11ConnectionHandler.processConnection(Http11BaseProtocol.java:664)
    11:47:08,468 ERROR [STDERR] at org.apache.tomcat.util.net.PoolTcpEndpoint.pr
    ocessSocket(PoolTcpEndpoint.java:527)
    11:47:08,468 ERROR [STDERR] at org.apache.tomcat.util.net.MasterSlaveWorkerT
    hread.run(MasterSlaveWorkerThread.java:112)
    11:47:08,468 ERROR [STDERR] at java.lang.Thread.run(Thread.java:595)
    # An unexpected error has been detected by HotSpot Virtual Machine:
    # EXCEPTION_ACCESS_VIOLATION (0xc0000005) at pc=0x29693a05, pid=832, tid=3776
    # Java VM: Java HotSpot(TM) Client VM (1.5.0_06-b05 mixed mode)
    # Problematic frame:
    # C [libdb44.dll+0x33a05]
    # An error report file with more information is saved as hs_err_pid832.log
    # If you would like to submit a bug report, please visit:
    # http://java.sun.com/webapps/bugreport/crash.jsp
    Press any key to continue . . .
    Thnking you
    Thanks & Regards
    Venkat Gadarla

    Hello,
    Your question is about Berkeley DB -- this is the Berkeley DB Java Edition forum. Please repost this over at:
    Berkeley DB
    Thanks.
    Charles Lamb

  • Problem while uploading master data through two files

    Hi all,
    When i try to upload master data of CUSTOMER with two files , Attribute and Text.
    Text gets uploaded perfectly.
    for Attribute, the data upload is done and psa i can see the exact data. But when i check the CUSTOMER object, i am not able to find the actually Attributs data. Instead those colmns are filed with 'E'.
    Can anybdy please explain me wheret the problem is.
    Thanks..

    Hi Sun,
    First check the data in PSA whether the records are fine or any discrepancy in them. Also check the properties of the COSTUMER object into which you are loading. The properties "type" or the "size" of the object may not be matching with the same in the file.
    Check them and try to load once again.
    Hope this helps u...
    Regards,
    Koundinya Karanam.
    Edited by: koundinya karanam on Jan 15, 2008 12:05 AM

  • Duplicate the details section when I retrieve data from Delivery notes

    Hey all.
    I built a query that retrieve data from ODLN and it works fine.
    But, when I have a Delivery Notes Document that has items from some different orders (that I copy them to Delivery notes of course) it duplicate my rows in the report
    I run the query at SQL server and it's work fine (on the same document)
    Is someone knows why this is happen?

    If one of these programs will not kick the iPod out of recovery mode than you are out of luck.
    For PC
    RecBoot: Easy Way to Put iPhone into Recovery Mode
    or
    http://joshuabailey1997.wordpress.com/2010/09/02/recboot-v1-3/
    If necessary:
    Download QTMLClient.dll & iTunesMobileDevice.dll for RecBoot
    and                                           
    RecBoot tip
    RecBoot may have problems on 64X windows, then try:       
    Tenorshare ReiBoot – Enter & Exit iPhone, iPad, iPod Recovery Mode with a Single Click
    For MAC or PC       
    The Firmware Umbrella - TinyUmbrella
    Installs blootware on PC too
    Your only home is to restore from your last backup
    iOS: Back up and restore your iOS device with iCloud or iTunes       

  • Problem importing specific data from .csv file

    Hello!
    I'm using JDev 11.1.3.0 with JSF ui and I've been following the excellent example about importing comma separated values from a .csv file, of Mr. Bors, but I have a specific problem...
    1) When i try to import a date value (format dd/mm/yyyy) to a date field of my adf table, the date won't be accepted and it's not shown (I'm using Row.setAttribute("Field", textValue) for that purpose). I guess I have to enter it in a specific format but don't know how!
    2) Same thing with double values. Tried 1.50 and it's showing on the adf table 0.02. Tried 1,50 (with a different text separator) and it's throwing me an error and doesn't show anything on the table (again i'm using Row.setAttribute("Field", textValue) for that one too)...
    Any help would be appreciated!
    Thank you
    Edited by: Nikolas Saridakis on 6 Νοε 2010 11:10 πμ

    I tried your suggestion before I set my question but didn't know if I was looking on the right direction (I tried some things but they didn't work out). After a bit poking everything was fine...
    So I used
    Row.setAttribute("Field_name", oracle.jbo.domain.Date.fromText(value, "dd/mm/yyyy", null));
    and it did the trick...
    Weird thing about numbers was that after that, they were entered properly on the table with a normal Row.setAttribute("Field_name", value); with value being a String... anyway!
    Thank you Timo!

Maybe you are looking for

  • Certified Flash Developer in CS3 or CS4?

    Back in 2004 it was possible to become a Macromedia Flash Certified Developer. No I wonder if it´s possible to become a certified developer in Flash CS3 or CS4. Can I do this online? Is there anyone who has the experience of this. I really appreciate

  • Create accessible PDFs that only open in Adobe Reader

    Is there a way to create/tag/mark PDFs so that they are forced to use Adobe Reader to open?  Sounds like a stupid question I know but we have a large number of accessible fillable PDF files that are now being opened by the integrated PDF readers in s

  • Lenovo G480 Problem: I cant Create a new Drive

    Hi everyone.. so my my laptop is a Lenovo G480 intel i3 4GB ram My operating System when i bought this was Windows 7 Home Basic. So my problem is that my Local Disk C is 420 GB while my Local Disk D is only 25.4 GB So I want my disk D to have more sp

  • Question About Xerces Parser and Java  JAXP

    Hi, I have confusion about both of these API Xerces Parser and Java JAXP , Please tell me both are used for same purpose like parsing xml document and one is by Apache and one is by sun ? And both can parse in SAX, or DOM model ? Is there any differe

  • FILE /usr/sap/trans/bin/DOMAIN.CFG could not be created

    Hi   After a DB copy into the QAS system the TMS will not connect to QAS. I set the CHGR3SHLOC to the domain controller and ran SE06 to delete and reset the TMS as DDIC in client 000. When I run TMS on the QAS box and try to include QAS in the DOMAIN