Issue regarding folder access from job server?

Hello,
We are using SAP BODS 4.0 sp2 setup. There is a requirement in our project. We are generating a text file as output and we are placing it in folder(particular path) on the job server. We have created user id and given access to the folder to end user.
Since the server is accessible through vpn. Is there any other way of sharing tat folder so that they can access in whatever network they are working.
The server is at Hyderabad and end users at Hyderabad can access through their LAN. But few users are some other places. How to provide access to them or enabling output file accessible to them.
Kindly let me know if any clarifications needed?
Regards,
karthik

Hi,
This is a network related question . You could do many things like creating a shared folder and shared across the network drive or write a .bat script to schedule to copy the file and put it in end user folder. Your network admin would be in a position to guide you better.
Arun

Similar Messages

  • Configure ODBC to access a CSV file on Linux for access from BI Server

    How to configure an ODBC connection to a CSV file on Linux for access from the BI Server Repository physical layer
    I am migrating a working windows OBIEE installation to Linux and can not seem to connect to csv files on Linux (from th BI server also running on the same Linux machine).
    I am using SUse Linux Enterprise server / 10 (Slash 10) standard odbc drivers
    My odbc ini file entries are:
    [ODBC]
    Trace=0
    TraceFile=odbctrace.out
    TraceDll=/app/oracle/product/10.1.3/OracleBI/odbc/lib/odbctrac.so
    InstallDir=/app/oracle/product/10.1.3/OracleBI/odbc
    UseCursorLib=0
    IANAAppCodePage=4
    [ODBC Data Sources]
    AnalyticsWeb=Oracle BI Server
    Cluster=Oracle BI Server
    SSL_Sample=Oracle BI Server
    idcbicsvfiles=Odbc Text driver
    [idcbicsvfiles]
    Description = Odbc Text driver
    Driver = Odbc Text driver
    Directory = /data/oracle/OracleBIData/idc
    #ReadOnly = No
    #CaseSensitive = No
    #Catalog = No
    ColumnSeperator = ,
    #Trace = 1
    #Tracefile =/data/oracle/OracleBIData/odbctrace.txt
    #Username      = oracle
    #Password      = ''
    [AnalyticsWeb]
    Driver=/app/oracle/product/10.1.3/OracleBI/server/Bin/libnqsodbc.so
    Description=Oracle BI Server
    ServerMachine=local
    Repository=
    Catalog=
    UID=
    PWD=
    Port=9703
    The csv files I want to use are in the directory /data/oracle/OracleBIData/idc to which I have set up a working and checked connection ([idcbicsvfiles]) on the linux machine itself.
    The error message I get when I select view data in the physcial layer is:
    [NQODBC][SQL_STATE: HY000][nQSError: 10058] A general error has occurred.
    [nQSError: 43093] An error occurred while processing the EXECUTE PHYSICAL statement.
    [nQSError: 16023] The ODBC function has returned ans error. The database may not be available, or the network may be down.
    Please can anybody give me a clue on how to get this working e.g. a working odbc.ini file from your own installation (and/or a tip for the odbc driver choice/configuration)
    P.S. I know this is not supported by Oracle but can not imagine, that nobody is using this.

    Hi,
    Chekc this...Re: Is there ODBC driver for excel flat file in Unix Box
    Re: OBIEE to use a CSV file as a data source on Linux
    Regards,
    Srikanth
    Edited by: Srikanth Mandadi on Oct 8, 2010 2:50 AM

  • WAD 7.0 issue regarding opening document from MIME repository

    Hi Experts,
    I am facing an issue in WAD 7.0 regarding opening document from MIME repository..
    My requirement
    I want to have a button , on clicking of this button one excel file ( already located at MIME Repository) should open in a new window.
    To achieve the same, I have added one button. and also added one java script web item in my web template.
    And as per my understanding, I need to write command/code in the script web item.
    I have tried with several standard commands available, but not yet succeeded.
    Please suggest the appropriate command or code , that I can put into the java script web item.
    Also please provide me, if there is any other solution.
    Thanks and Regards,
    Biswarup Ghosh

    Yes I tried it at first with Script Item but it did not work. I am not sure how should I change the path because of the portal:
    "/sap/bw/Mime/Customer/.../file.htm" - this path is not working anymore
    Another option that you mentioned I do not want to use. I have about 20 different html files in MIME repository so I would have to create 20 different web application. I do not like this solution.
    Regards
    Erwin

  • Migration issue in SQL Developer (from SQL Server to Oracle)

    Dear All,
    I am using SQL Developer 3.0 to migrate data from SQL Server 2008 to Oracle 11gR2. I used Quick Migration Wizard for migrating and everything was Ok but when it was running migration process, I got this message: “Migration actions have failed. Check the migration reports for details. : null”
    But there was not any report to check. Does anyone know how I can do it without any problem?
    Thanks in advance.

    I am working on a project to migrate from SQL Server 2008 to Oracle 11gR2. It includes some tables, views, indexes, triggers and store procedures. At first I worked on SQL Developer 2.1.1.64 but it doesn’t support SQL Server 2008 (I had to use SQL Server 2005 in its offline options). I got a lot of error when I used the offline options. I know that I may be having some problem in table or view definition in my SQL Server but if I don’t have any log file I can’t solve them.
    Unfortunately, in SQL Developer 3.0 there isn’t any option to generate a SQL comments from objects (convert to Oracle) because I need to rewrite comments if they have problem. For example I have a filed named USER that it’s a reserved word in Oracle so I have to change this field name.
    I also can’t use “Translation Scratch Editor” because it has some error to convert. For example when I use it for the following code:
    CREATE TABLE [dbo].[Table1](
         [ID1] [int] IDENTITY(1,1) NOT NULL,
         [Name1] [varchar](50) NULL,
         CONSTRAINT [PK_Table1] PRIMARY KEY CLUSTERED
         [ID1] ASC
    )WITH (PAD_INDEX = OFF, STATISTICS_NORECOMPUTE = OFF, IGNORE_DUP_KEY = OFF, ALLOW_ROW_LOCKS = ON, ALLOW_PAGE_LOCKS = ON) ON [PRIMARY]
    ) ON [PRIMARY]
    CREATE TABLE [dbo].[Table2](
         [ID2] [int] IDENTITY(1,1) NOT NULL,
    [Name1] [varchar](50) NULL,
         CONSTRAINT [PK_Table2] PRIMARY KEY CLUSTERED
         [ID2] ASC
    )WITH (PAD_INDEX = OFF, STATISTICS_NORECOMPUTE = OFF, IGNORE_DUP_KEY = OFF, ALLOW_ROW_LOCKS = ON, ALLOW_PAGE_LOCKS = ON) ON [PRIMARY]
    ) ON [PRIMARY]
    I receive a wrong code like this:
    CREATE SEQUENCE Table1_ID1
    START WITH 1
    INCREMENT BY 1;
    CREATE TABLE Table1
    ID1 NUMBER(10,0) NOT NULL,
    Name1 VARCHAR2(50) ,
    CONSTRAINT PK_Table1 PRIMARY KEY( ID1 )
    CREATE OR REPLACE TRIGGER Table1_ID1_TRG
    BEFORE INSERT
    ON Table1
    FOR EACH ROW
    BEGIN
    SELECT Table1_ID1.NEXTVAL INTO :NEW.ID1
    FROM DUAL;
    END;
    CREATE SEQUENCE Table1_ID1
    START WITH 1
    INCREMENT BY 1;
    CREATE SEQUENCE Table2_ID2
    START WITH 1
    INCREMENT BY 1;
    CREATE TABLE Table2
    ID2 NUMBER(10,0) NOT NULL,
    Name1 VARCHAR2(50) ,
    CONSTRAINT PK_Table2 PRIMARY KEY( ID2 )
    CREATE OR REPLACE TRIGGER Table1_ID1_TRG
    BEFORE INSERT
    ON Table1
    FOR EACH ROW
    BEGIN
    SELECT Table1_ID1.NEXTVAL INTO :NEW.ID1
    FROM DUAL;
    ENDCREATE OR REPLACE TRIGGER Table2_ID2_TRG
    BEFORE INSERT
    ON Table2
    FOR EACH ROW
    BEGIN
    SELECT Table2_ID2.NEXTVAL INTO :NEW.ID2
    FROM DUAL;
    END;

  • Issue with migration DB from SQL Server 2005 to Oracle 11.2 using SQL Developer Migration workbench

    Hi,
    We face an issue while migrating an SQL Server 2005 DB to Oracle 11.2.  It fails during the process.  I hope someone on the forum has seen this before and can give us some advice.
    I use the latest version of SQL Developer with JRE included (3.2.20.09).  The JDBC driver to connect to SQL Server 2005 is 1.2.0
    Here are the steps we take in the Migration Workbench wizard:
    I made an online extract of the SQL Server database.
    Using the workbench in SQL Developer I created a migration repository in schema PDM_MIGRATION.
    Next I start the migration, it captures the tables fine, but immediately after the start of the conversion I get a failed message without further explanations.
    This is the content of the error xml:
    <?xml version="1.0" encoding="windows-1252" standalone="no"?>
    <log>
    <record>
      <date>2013-08-14T16:23:32</date>
      <logger>oracle.dbtools.migration.workbench.core.MigrationLogResourceBundle</logger>
      <level>SEVERE</level>
      <class>oracle.dbtools.migration.workbench.core.logging.MigrationLogUtil</class>
      <message>Ongeldig naampatroon.: PDM_MIGRATION .MIGR_FILTER</message>
      <param>oracle.dbtools.migration.workbench.core.logging.LogInfo@4c12ab</param>
      <exception>
        <message>oracle.dbtools.migration.convert.ConvertException: Ongeldig naampatroon.: PDM_MIGRATION .MIGR_FILTER</message>
        <frame>
          <class>oracle.dbtools.migration.convert.ConverterWorker</class>
          <line>1078</line>
        </frame>
        <frame>
          <class>oracle.dbtools.migration.convert.ConverterWorker</class>
          <line>316</line>
        </frame>
        <frame>
          <class>oracle.dbtools.migration.workbench.core.ui.FullMigrateTask</class>
          <line>1002</line>
        </frame>
        <frame>
          <class>oracle.dbtools.migration.workbench.core.ui.FullMigrateTask</class>
          <line>303</line>
        </frame>
        <frame>
          <class>oracle.dbtools.migration.workbench.core.ui.FullMigrateTask</class>
          <line>205</line>
        </frame>
        <frame>
          <class>oracle.dbtools.migration.workbench.core.ui.FullMigrateTask</class>
          <line>159</line>
        </frame>
        <frame>
          <class>oracle.dbtools.raptor.backgroundTask.RaptorTask</class>
          <line>193</line>
        </frame>
        <frame>
          <class>java.util.concurrent.FutureTask$Sync</class>
          <line>303</line>
        </frame>
        <frame>
          <class>java.util.concurrent.FutureTask</class>
          <line>138</line>
        </frame>
        <frame>
          <class>oracle.dbtools.raptor.backgroundTask.RaptorTaskManager$RaptorFutureTask</class>
          <line>515</line>
        </frame>
        <frame>
          <class>java.util.concurrent.Executors$RunnableAdapter</class>
          <line>441</line>
        </frame>
        <frame>
          <class>java.util.concurrent.FutureTask$Sync</class>
          <line>303</line>
        </frame>
        <frame>
          <class>java.util.concurrent.FutureTask</class>
          <line>138</line>
        </frame>
        <frame>
          <class>java.util.concurrent.ThreadPoolExecutor$Worker</class>
          <line>886</line>
        </frame>
        <frame>
          <class>java.util.concurrent.ThreadPoolExecutor$Worker</class>
          <line>908</line>
        </frame>
        <frame>
          <class>java.lang.Thread</class>
          <line>662</line>
        </frame>
      </exception>
    </record>
    <record>
      <date>2013-08-14T16:23:32</date>
      <logger>oracle.dbtools.migration.workbench.core.MigrationLogResourceBundle</logger>
      <level>SEVERE</level>
      <class>oracle.dbtools.migration.workbench.core.logging.MigrationLogUtil</class>
      <message>Ongeldig naampatroon.: PDM_MIGRATION .MIGR_FILTER</message>
      <param>oracle.dbtools.migration.convert.ConverterWorker.copyModel(ConverterWorker.java:1078)</param>
      <param>oracle.dbtools.migration.convert.ConverterWorker.runConvert(ConverterWorker.java:316)</param>
      <param>oracle.dbtools.migration.workbench.core.ui.FullMigrateTask.doConvert(FullMigrateTask.java:1002)</param>
      <param>oracle.dbtools.migration.workbench.core.ui.FullMigrateTask.doMaskBasedActions(FullMigrateTask.java:303)</param>
      <param>oracle.dbtools.migration.workbench.core.ui.FullMigrateTask.doWork(FullMigrateTask.java:205)</param>
      <param>oracle.dbtools.migration.workbench.core.ui.FullMigrateTask.doWork(FullMigrateTask.java:159)</param>
      <param>oracle.dbtools.raptor.backgroundTask.RaptorTask.call(RaptorTask.java:193)</param>
      <param>java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303)</param>
      <param>java.util.concurrent.FutureTask.run(FutureTask.java:138)</param>
      <param>oracle.dbtools.raptor.backgroundTask.RaptorTaskManager$RaptorFutureTask.run(RaptorTaskManager.java:515)</param>
      <param>java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:441)</param>
      <param>java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303)</param>
      <param>java.util.concurrent.FutureTask.run(FutureTask.java:138)</param>
      <param>java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886)</param>
      <param>java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908)</param>
      <param>java.lang.Thread.run(Thread.java:662)</param>
      <param>oracle.dbtools.migration.workbench.core.logging.LogInfo@5dc1bc</param>
      <exception>
        <message>oracle.dbtools.migration.convert.ConvertException: Ongeldig naampatroon.: PDM_MIGRATION .MIGR_FILTER</message>
        <frame>
          <class>oracle.dbtools.migration.convert.ConverterWorker</class>
          <line>1078</line>
        </frame>
        <frame>
          <class>oracle.dbtools.migration.convert.ConverterWorker</class>
          <line>316</line>
        </frame>
        <frame>
          <class>oracle.dbtools.migration.workbench.core.ui.FullMigrateTask</class>
          <line>1002</line>
        </frame>
        <frame>
          <class>oracle.dbtools.migration.workbench.core.ui.FullMigrateTask</class>
          <line>303</line>
        </frame>
        <frame>
          <class>oracle.dbtools.migration.workbench.core.ui.FullMigrateTask</class>
          <line>205</line>
        </frame>
        <frame>
          <class>oracle.dbtools.migration.workbench.core.ui.FullMigrateTask</class>
          <line>159</line>
        </frame>
        <frame>
          <class>oracle.dbtools.raptor.backgroundTask.RaptorTask</class>
          <line>193</line>
        </frame>
        <frame>
          <class>java.util.concurrent.FutureTask$Sync</class>
          <line>303</line>
        </frame>
        <frame>
          <class>java.util.concurrent.FutureTask</class>
          <line>138</line>
        </frame>
        <frame>
          <class>oracle.dbtools.raptor.backgroundTask.RaptorTaskManager$RaptorFutureTask</class>
          <line>515</line>
        </frame>
        <frame>
          <class>java.util.concurrent.Executors$RunnableAdapter</class>
          <line>441</line>
        </frame>
        <frame>
          <class>java.util.concurrent.FutureTask$Sync</class>
          <line>303</line>
        </frame>
        <frame>
          <class>java.util.concurrent.FutureTask</class>
          <line>138</line>
        </frame>
        <frame>
          <class>java.util.concurrent.ThreadPoolExecutor$Worker</class>
          <line>886</line>
        </frame>
        <frame>
          <class>java.util.concurrent.ThreadPoolExecutor$Worker</class>
          <line>908</line>
        </frame>
        <frame>
          <class>java.lang.Thread</class>
          <line>662</line>
        </frame>
      </exception>
    </record>
    <record>
      <date>2013-08-14T16:23:32</date>
      <logger>oracle.dbtools.migration.workbench.core.MigrationLogResourceBundle</logger>
      <level>WARNING</level>
      <class>oracle.dbtools.migration.workbench.core.ui.FullMigrateTask</class>
      <message>Building converted model: FAILED : Database Migration : FAILED</message>
      <param>oracle.dbtools.migration.workbench.core.logging.LogInfo@15a3779</param>
    </record>
    Does anybody know what this error means and what steps we should take to continue the migration?
    I see the PDM_MIGRATION.MIGR_FILTER is a type.
    Many thanks in advance,
    Kris

    Hi Wolfgang,
    Thanks for your reply.
    This is how the type MIGR_FILTER looks like:
    create or replace
    TYPE MIGR_FILTER IS OBJECT (
      FILTER_TYPE INTEGER, -- Filter Types are 0-> ALL, 1->NAMELIST, 2->WHERE CLAUSE, 3->OBJECTID LIST
      OBJTYPE VARCHAR2(40),
      OBJECTIDS OBJECTIDLIST,
      NAMES NAMELIST,
      WHERECLAUSE VARCHAR2(1000));
    I think the repository user has the correct privileges.  This is the overview of privileges it has:
    SQL> select * from dba_sys_privs where GRANTEE in ('PDM_MIGRATION') order by GRANTEE;
    GRANTEE PRIVILEGE ADM
    PDM_MIGRATION ALTER SESSION NO
    PDM_MIGRATION CREATE CLUSTER NO
    PDM_MIGRATION CREATE DATABASE LINK NO
    PDM_MIGRATION CREATE PROCEDURE NO
    PDM_MIGRATION CREATE SEQUENCE NO
    PDM_MIGRATION CREATE SESSION NO
    PDM_MIGRATION CREATE SYNONYM NO
    PDM_MIGRATION CREATE TABLE NO
    PDM_MIGRATION CREATE TRIGGER NO
    PDM_MIGRATION CREATE VIEW NO
    PDM_MIGRATION UNLIMITED TABLESPACE NO
    SQL> select * from dba_role_privs where GRANTEE in ('PDM_MIGRATION') order by GRANTEE;
    GRANTEE GRANTED_ROLE ADM DEF
    PDM_MIGRATION CONNECT NO  YES
    PDM_MIGRATION RESOURCE NO  YES
    Best regards,
    Kris

  • File Access from DB Server/Unix Mounting

    We have an Oracle Database that runs on Unix Box and a cold fusion application/web server runs on a Windows box. We have a data intensive stored procedure that gets data from DB, does lot of massaging and then create bunch of html files with sorting capabilities etc. This process happens upon user request from website (cold fusion/html page calling Oracle stored procedure). After this, the cold fusion server need to get access to these html files to display to the user.
    Our Unix DBA's does NOT want to do File Mounting as they think the windows box is not secure. Hence we are exploring other options.
    Below are the options we are exploring. But very well open to suggestions...
    1. Let cold fusion query the data from Database and let it do the data manipulation/massaging
    2. We explored the external tables (based on file) but the files are dynamic and could not find a way to do this
    3. FTP transfer from Unix to NT box but there could be time delay and may not be available for immediate display
    4. Create a temporary table, let the DB procedure insert data into table. Then Cold fusion will query from there.
    Any other suggestions?

    Hi,
    Its not required to handle for #
    just read the dataset into the string and split the string into work area fields seoparated by TAb delimiter.
    Eg:
    READ DATASET <filename> INTO l_string.
    SPLIT l_string
                   AT cl_abap_char_utilities=>horizontal_tab "Horizontal tab
                  INTO l_funcloc_clas-tplnr
                       w_funcloc_clas-class
                       w_funcloc_clas-klart
                       w_funcloc_clas-atnam
                       w_funcloc_clas-atwrt
                       w_funcloc_clas-atfor.
    append w_funcloc_clas to l_funcloc_clas.
    With Regards,
    Dwarakanath.S

  • Issue with Teradata access from Business Objects AIX Machine

    Hi,
    I am using BO  XI.31 which is installed on AIX 5.3 machine. I have installed Teradata driver on the machine and I have created the require ODBC.ini entry to point Teradata database. While refreshing the report I am getting below error message.
    A database error occured .The database error test is (CS) "DBDriver failed ot load: /BOBJ/xi_3.1...Connectionserver/libdbd_tdodbc.so
    0509-022 Cannot load module/BOBJ/xi_3.1....ConnectionServer/libdbd_tdodbc.so
    0509-150 Dependent module libodbc.a(odbc.so) Could not be loaded.
    0509-022 Cannot load module libodbc.a(odbc.so)
    0509-022 System error:A file or directory in the path name does not exist.
    Any body know what are the steps require to access teradata database from AIX box

    Hi Jacks,
    Following information might help you to resolve the issue.
    This is caused by a wrong HOME environment variable of the user that starts the BO processes.
    Resolution
    Check the correct path of your BO 3.1 installation and export it as HOME environment variable:
    Example:
    BO 3.1 installation path: /home/<bouser>/bo3.1/Enterprise12.0../...
    Type in the command line:
    export HOME=/home/<bouser>/boversion
    You can also change this environment variable permanently if you edit the .profile file of the BO user (home/<username>/.profile), adding the above line into the file, then source it again with:
    . .profile
    or with a logoff/logon of the session.
    Check if the path of your HOME is correct with:
    env | grep HOME
    Now you can run the installation setup without any error.
    Regards,
    Sarbhjeet Kaur

  • Issue with welcome prompt from VXML server

    Hi
    I am new to Contact Center. In lab i have build sprawller and CVP ( comprehensive).
    I have a small script which should play welcome promt, but i am having issues to play the "Welcome" prompt. The call hits the sccript but the welcome message is not played.
    Please help me to fix the issue.
    Debugs from http client on VXML gateway.
    *Mar  1 16:42:28.723: //331//HTTPC:/httpc_cache_entry_free: cache(650AD474)URL:http://192.168.0.41:8000/cvp/VBServlet?MSG_TYPE=PING&CALL_DNIS=140198&CALL_ANI=sip:[email protected]:5060&ERROR_CODE=0&RECOVERY_VXML=flash:recovery.vxml&CLIENT_TYPE=IOS&CALL_ID=119BABC91000012A4A50D9073F57FFD7&SIP_CALL_ID=[email protected]&CALL_UUI=&VERSION=7.0.2
    *Mar  1 16:42:31.639: //331//HTTPC:/httpc_cache_entry_free: cache(68123DFC)URL:http://192.168.0.41:8000/cvp/VBServlet?MSG_TYPE=CALL_NEW&CALL_DNIS=140198&CALL_UUI=&CALL_ANI=sip:[email protected]:5060&RECOVERY_VXML=flash:recovery.vxml&CLIENT_TYPE=IOS&CALL_ID=119BABC91000012A4A50D9073F57FFD7&ERROR_CODE=0
    *Mar  1 16:42:35.887: //331//HTTPC:/httpc_cache_entry_free: cache(66797988)URL:http://192.168.0.41:7000/CVP/en-us/app/Welcome.wav
    *Mar  1 16:42:40.079: //331//HTTPC:/httpc_cache_entry_free: cache(650AC40C)URL:http://192.168.0.41:8000/cvp/VBServlet?MSG_TYPE=CALL_RESULT&CALL_ID=119BABC91000012A4A50D9073F57FFD7&CALL_SEQ_NUM=1&ERROR_CODE=0
    *Mar  1 16:42:42.875: //331//AFW_:/vapp_stop_fetchaudio_timer: No active leg
    *Mar  1 16:42:43.035: //331//HTTPC:/httpc_cache_entry_free: cache(6679801C)URL:http://192.168.0.41:8000/cvp/VBServlet?MSG_TYPE=CALL_RESULT&CALL_ID=119BABC91000012A4A50D9073F57FFD7&CALL_SEQ_NUM=2&ERROR_CODE=1

    Something looks really strange in that trace. I don't see  the IVR transfer label coming over. Do you have a "Send To VRU" node?
    What does this script look like? Here is a simple example.
    You should start with a simple play media microapp.
    If using the CVP VXML server, you should first test using the supplied HelloWorld application. That CVP application is already deployed in the VXML server - just call it correctly. The nodes above need to be a little different.
    PS. The log file is full of Reporting rubbish. Since you don't have the Reporting Server and DB, tell the Call Server there is no reporting. It will clean up the deadwood.
    Regards,
    Geoff

  • Data access from Application Server - Seeking Opinion

    I am working on a fairly large scale ERP application that is written in Java both on the front end, and middle tier ( using a JBoss application server ).
    All of the database access happens in my app server, and when I wanted to get peoples opinions on the best way to extract and pass data to and from the database from the app server, and of course to and from the client.
    I am using JDBC to make database calls and extra data from the database. This is of course pretty trivial. And I first started writing POJO that represent tables in my database. I then wrote a faily length method that uses reflection to call the setting methods and pass in the objects returned from the database result set. This of course becomes tricky when you start to join multiple tables together. And often you only need a small subset of columns in a table and don't require an object with all the columns from the table ( some set, and some as null objects ).
    So then I decided I was use a combination of ArrayList objects and HashMap's to store the data. If a result set returned multiple rows, it would return an ArrayList of HashMap objects. Each hash map would contain the data for that row, and the hash map keys are the column names.
    This seems to work pretty well and resolves the problem of joining multiple tables together and keeping track of which columns get stored in which objects.
    Does anyone have a different solution or idea as to how to handle this? And thoughts or ideas would be greatly appreciated.

    bryano wrote:
    Let me pose a quick Hibernate question as well.Let me recommend that you not be so thin-skinned about responses.
    >
    If you had a table that had 40 columns in it and say 1000 rows. And you needed to run a query that returned all 1000 rows, but you only needed two out of the
    40 columns. Would it be better to extract those two columns into the HashMap / ArrayList collection I mentioned in my original post? Or would using something like Hibernate and a class that mapped all of the columns in the table be okay?Why not just map the columns you needed in Hibernate? Who said you had to map all 40 and have them be null?
    My concern is on efficiency,
    and I was wondering if building 1000 objects that each have 40 members that only 2 are populated is the most efficient way of extracting the data.Doesn't sound very efficient.
    I will admit, my knowledge of Hibernate is limited at best so I may be missing a component of Hibernate that would allow you to only extract the columns you required, but you are still working with an object that has 38 null value objects for the columns you didn't require.I don't believe you're required to map every column in a table.
    %

  • Validating file/folder path from application server.

    Hi All ,
    I have used the FM '/SAPDMC/LSM_F4_SERVER_FILE' to open F4 help for application server. Now I want to validate the folder or file returned by the FM.
    Can anyone help on this?
    Thanks
    Shilpa

    HI Shilpa,
    Us this  FM
    F4_DXFILENAME_TOPRECURSION
    Hope this will help you.
    Regards,
    Vijay

  • Laserjet CP2025 share issue (64-bit drivers from Windows Server 2003 32-bit).

    Hi,
    We are migrating our Desktops from XP 32-bit to Windows 7 64-Bit and I'm trying to update all of our printer drivers on our Print Server (Windows server 2003 32 Bit).
    I am adding the new drivers by opening up each printers properties on the print server, going to Sharing and then add new driver and directing it to the relevant folder for the 64-bit drivers for the printer.
    Every printer (HP, Panasonic, Brother) has worked so far, except for the HP Laserjet CP2025. The Net Express installer doesn't work. Nor do the Universal Printing Drivers (PLC 6). Does anyone know a way around this?
    Thank you.

    Can anyone help with this? I'm in a difficult position with this. Any help would be appreciated.

  • Transparency issue when generating PDF from InDesign Server

    Hi there
    Has anyone experienced any issues with transparency appearing solid white when generating a PDF from the InDesign server? There is no issue when exporting from the InDesign application, only when using the Server.
    Any pointers / suggestions would be much appreciated
    Thanks
    Davey

    Davey, please send me an InDesign Server document that I could use to reproduce this problem. My e-mail address is [email protected] Thank you. --DougR

  • Folder Access on a server

    I want to create a folder on our server that allows access to some, but restricts others. Can I edit or create a 'Group' under 'Details' in the Info box, or is that created solely by the OS?
    If not, what's the best way to accomplish this?
    Thanks!
    OSX Server   Mac OS X (10.4.9)  
      Mac OS X (10.4.9)  

    You'll have to ask the Server Admin to create a new workgroup that contains all of the users with whom you want to share the folder. The group should include you. Then you yourself should be able to set the group privilege on any folder you own.

  • [ACS 5.4] Logs access from secondary server

    Hi,
    I have 2 ACS 5.4 in distributed environment. Everything left to defaults besides policy.
    Let assume ACS-A is the primary and ACS-B is the secondary. Regularly, I'd connect to ACS-A to make changes and WATCH LOGs.
    Now, let assume ACS-A is down. Obviously, I connect to ACS-B and everything works fine, besides logs. When I click on 'logs center', a blank window opens and nothing happens.
    But the URL it tries to open, it's ACS-A.
    Now, from what I saw, ACS-A being the primary box is the log collector for a distributed environment, by default. But how I supposed to watch the logs on a secondary server when primary is down?
    Thank you.

    Hello Alex,
    The following are the supported browsers and it should work fine in all fo them. Please have a look at them:-
    Supported Web Client and Browsers
    You can access the ACS 5.4 administrative user interface using the following web clients and browsers:
    •MAC Platform
    –Mozilla Firefox version 3.x
    –Mozilla Firefox version 10.x
    •Windows 7 32-bit
    •Windows 7 64-bit
    •Windows XP Professional (Service Pack 2 and 3)
    –Internet Explorer version 7.x
    –Internet Explorer version 8.x
    –Internet Explorer version 9.x
    –Mozilla Firefox version 3.x
    –Mozilla Firefox version 8.x
    –Mozilla Firefox version 9.x
    –Mozilla Firefox version 10.x
    The above mentioned browsers are supported only with one of the following cipher suits:
    –-TLS_RSA_WITH_AES_256_CBC_SHA
    –-TLS_RSA_WITH_AES_128_CBC_SHA
    –-RSA_WITH_3DES_EDE_CBC_SHA

  • Regarding error while scheduling Job Server (rapid mart)

    hi friends
    iam new to bods we are using BODS 3.2 source r/3 and target is MYSQLSERVER . and we have installed rapidmarts
    General_ledger_rapid mart during scheduling the job fails and got and
    erorr
    Transform EID_2_Rename_ConvertDateToChar Cannot convert string <1900.01.01 00:00:00> to date using format string <YYYY-MM-DD>.
    (12.2) 02-11-12 04:59:09 (W) (1124:11052) RUN-053006: |Data flow DF_FinDocFactNewGL_SAP|Embedded Dataflow Copy_1_EDF_FinDocFact_SAP|Transform EID_2_Rename_ConvertDateToChar
    please help how to rectify this step by step ( as iam new )
    please

    thanks suneer
    can you please explain if i find the problem and rectify
    and after that how could i schedule the jobs ( as it is a rapid mart it contains number of data flow) so,
    how should i schedule or do i need to schedule this DF only in the designer or through Data Service Management console
    I think if i schedule it will again schedule all the jobs in the rapid mart ( i want to schedule only one DF which is eror so, how should i do)
    and if it is"? it will show duplicate reports
    please suggest
    thanks

Maybe you are looking for

  • Simple help: on a mac

    Ok guys, I have a very simple question...but can't seem to find an answer. I there anyway I can download flash for my mac? I've found downloads for flash PLAYER on mac...but I don't really understand the difference. All I'm looking for is the version

  • SBS 2003 Server & email

    Hi everyone! This is my third post regarding not being able to set up my email and I found out a little more info today. We are using SBS 2003 for a server and this is maybe why I'm having problems. We use excange and access our email with Microsoft

  • 4g lte

    Phone selling in india is showing that it had facilitated with 4g features. But right now 4g is not made available by networks in india. Any way it is about to launch in year 2015 which is short near by 6 monts. I want to know wether my z1 compact bu

  • Epson Perfection "Supported Twain Driver"

    I moved my Epson Perfection 1670 over to my new Apple Mini Server, installed the Epson Software, and get this message when trying to scan:  "Please install the supported Twain driver."  I have looked everywhere on Epson's web site for the "Supported

  • Shipping updates for 10/3

    Has anyone received shipping updates for their iphone 6 or 6 plus with the expected ship date of 10/3? Post your updates here!