BODS Vs BW

Hi All ,
Looking for capabilties between BODS Vs SAP BW.
Possible more in details
like
Capabilty -   SAP BW                      Vs   BODS_
Extraction    Data Centric   -             Process Specific
Cleansing
Quality
Detla Handling
DB Extraction
XML Extraction
Web Data
Accesing Million Records (Avg Time Taken )
Let me know ,
This will helps to showcasing to the customer also.
Regards,
Mahesh

hi,
  kindly check the below link for BODS:
http://www.msatish.com/sap-businessobjects-sap-businessobjects-data-servicesbods/
http://www.sap.com/solutions/sapbusinessobjects/large/eim/data-services/index.epx
for BW features/benefits
http://www.sap.com/solutions/sapbusinessobjects/data-warehousing/sapnetweaver-business-warehouse/index.epx
since you will be presenting to client.. few cents from my side since the client would not be interested more in technical side he/she will be more interested in how he can get data/reports  asap at his end i.e. time/TCO and also in a cost effective manner(Money) and the amount data that can be handled by both tools  so if you give your presentation in that manner it gives clear cut idea idea you can optimize the client resources perfectly.
my suggestions just for your reference
regards
laksh

Similar Messages

  • Error while executing a uinx script in BODS job - exit code 127

    I am executing a UNIX script using BODS job and it is giving an error message saying 'Program terminated because of exit code 127'. The same script i am able to execute in UNIX. I am not able to find the solution for this.

    Using a print command with return code 8 should give more insight to the error.
    print(exec('sh \bin','<your command>',8));

  • BODS : Datastore options for SAP R/3 - need clarity for use

    Hi All.
    Another request to understand the datastore optiono n BOXI 3.1 BODS existing installation.
    We are trying to pull a new table from SAP R/3 into BODS and we find that ABAP program is not getting generated as expected. And terminates. When we tested a simple workflow.
    While creating a new datastore for SAP R/3 Source :
    When we look at the datastore we find the following options for R3 Source ;
    ABAP Execution option :
    Execute preloaded
    Generate and execute
    Under data transfer method we find ;
    Shared directory
    Direct download
    FTP
    Custom transfer
    Then we have working directory on sap server
    local directory
    generated ABAP Directory
    Am testing a simple workflow of pulling data from SAP R/3 in a dev machine.
    But am not understanding.
    which option of ABAP Execution, and data transfer , path would work in co-ordination.
    Because when i say direct download and say generate and execute it throws error.
    Can anyone help me with combinations of the options to choose for a R/3 source. And the implications thereof.
    I had created a folder under local bods server D:\Bodi. And given the path for data transfer.
    But the files are not getting generated for whatsoever reasons.
    Any advise on this would be helpful.
    Also found a bit unusual, that there was no button to test the connection to be correct or not, a TEST connection button is not there. Which i felt, could be included.
    Note : on the existing production system, we have chosen execute preloaded; shared directory on sap server. and shared that folder path for the user and given full rights. But while we try to do the same on the dev machine a test before transporting on production, a simple workflow does not work.
    would like to know what settings on sap server, really affect the data store options on the BODS
    thanks
    indu
    Edited by: Indumathy Narayanan on Jul 19, 2011 4:14 PM

    Indeed, BODS <> SAP connectivity can be tricky.
    For a development environment, I suggest you select the option "Generate and execute" for your "ABAP Execution Option." What this means is that DS will create, on the fly, small-ish ABAP programs. These ABAP programs will be written, in plain text, to a local directory on the DS job server, in the folder specified in "Generated ABAP Directory".  You can see them in there after an attempted job execution, assuming the job involves the creation of an ABAP program.  The ABAP program name is specified in the properties of an ABAP data flow, under Options > ABAP Program Name.  If you can't, perhaps the DS job server process doesn't have full rights to that folder - ?  After being generated on-the-fly, they'll be transferred to SAP to execute.  The SAP user you use to connect to SAP must have sufficient rights to upload-and-execute these ABAP programs, and that's a fairly substantial set of rights. What's required is documented in the BODS supplement for SAP. Often, to get things running, your friendly local Basis admin will grant SAP_ALL to the DS user, to see what rights are being invoked.
    Once all that jazz is working, you need to get the data back. There are a number of ways to do this. The method of data transfer is specified in "Data transfer method," where, ignoring "Custom transfer," you have three choices:
    1) Direct Download: easiest and slowest.  This method tells SAP to attempt to stick data in the client-side folder specified in "Local directory." Try this first.
    2) Shared folder: This is recommended when you have SAP being hosted on a Windows box. Basically: you set "Working directory on SAP Server" and "Data Services path to the shared directory" to point to the same folder.  SAP uses the "Working directory on SAP Server" to find this folder, and DS uses the other setting. So, for instance, if you were going to use the Shared folder method, you could set "Working directory on SAP Server" to "E:\BODS_Transfer", and, assuming E:\BODS_Transfer was shared-out as "BODS_Transfer", you could set "Data Services path to the shared directory" to
    dev12.somecompany.com\BODS_Transfer .  Then, you'd need to setup all the relevant security, as both SAP and DS need rights to read and write files in this folder.
    3) FTP (this is the method I usually use): SAP writes the "transport files" you're after (i.e., the data) in the folder specified in "Working directory on SAP Server". Then, you need to establish ftp connectivity to that folder from the DS job server's perspective, which you do by entering the ftp host name and the path to that folder in "FTP host name" and "FTP relative path to the SAP working directory".  In my opinion, the "relative" business is a little confusing, and I just typically enter the full ftp path, beginning the path w/ a forward slash, like "/usr/sap/tmp/BOBJ" or something like that.  You also need to obtain a separate username and password for the ftp connectivity. Note that this name and password has NOTHING to do with the SAP username and password; you're just setting-up DS to act as an ftp client. I strongly encourage you to test ftp connectivity by using a regular ftp client from the DS job server and attempt to connect to your ftp host using the username and password you were given, and attempting to fetch some sample test file. If you can't do this, manually, then DS won't be able to do it, either.
    Best wishes,
    Jeff Prenevost

  • BODS 3.1 : SAP R/3 data extraction -What is the difference in 2 dataflows?

    Hi.
    Can anyone advise as to what is the difference  in using the data extraction flow for extracting Data from SAP R/3 ?
    1)DF1 >> SAPR/3 (R3/Table -query transformation-dat file) >>query transformation >> target
    This ABAP flow generates a ABAP program and a dat file.
    We can also upload this program and run jobs as execute preloaded option on datastore.
    This works fine.
    2) We also can pull the SAP R/3 table directly.
    DF2>>SAPR/3 table (this has a red arrow like in OHD) >> Query transformation >> target
    THIS ALSO Works fine. And we are able to see the data directly into oracle.
    Which can also be scheduled on a job.
    BUT am unable to understand the purpose of using the different types of data extraction flows.
    When to use which type of flow for data extraction.
    Advantage / disadvantage - over the 2 data flows.
    What we are not understanding is that :
    if we can directly pull data from R/3 table directly thro a query transformation into the target table,
    why use the Flow of creating a R/3 data flow,
    and then do a query transformation again
    and then populate the target database?
    There might be some practical reasons for using these 2 different types of flows in doing the data extraction. Which I would like to understand.  Can anyone advise please.
    Many thanks
    indu
    Edited by: Indumathy Narayanan on Aug 22, 2011 3:25 PM

    Hi Jeff.
    Greetings. And many thanks for your response.
    Generally we pull the entire SAP R/3 table thro query transformation into oracle.
    For which we use R/3 data flow and the ABAP program, which we upload on the R/3 system
    so as to be able to use the option of Execute preloaded - and run the jobs.
    Since we do not have any control on our R/3 servers nor we have anyone on ABAP programming,
    we do not do anything at the SAP R/3 level
    I was doing this trial and error testing on our Worflows for our new requirement
    WF 1 : which has some 15 R/3 TABLES.
    For each table we have created a separate Dataflow.
    And finally in between in some dataflows, wherein, the SAP tables which had lot of rows, i decided to pull it directly,
    by-passing the ABAP flow.
    And still the entire work flow and data extraction happens ok.
    In fact i tried creating a new sample data flow and tested.
    Using direct download and - and also execute preloaded.
    I did not see any major difference in time taken for data extraction;
    Because anyhow we pull the entire Table, then choose whatever we want to bring into oracle thro a view for our BO reporting or aggregate and then bring data as a table for Universe consumption.
    Actually, I was looking at other options to avoid this ABAP generation - and the R/3 data flow because we are having problems on our dev and qa environments - giving delimiter errors.  Whereas in production it works fine. Production environment is a old set up of BODS 3.1. QA and Dev are relatively new enviornments of BODS. Which is having this delimiter error.
    I did not understand how to resolve it as per this post : https://cw.sdn.sap.com/cw/ideas/2596
    And trying to resolve this problem, I ended up with the option of trying to pull directly the R/3 table. Without using ABAP workflow.  Just by trial and error of each and every drag and drop option. Because we had to urgently do a POC and deliver the data for the entire e recruiting module of SAP. 
    I dont know whether i could do this direct pulling of data - for the new job which i have created,
    which has 2 workflows with 15 Dataflows in each worflow.
    And and push this job into production.
    And also whether i could by-pass this ABAP flow and do a direct pulling of R/3 data, in all the Dataflows in the future for ANY of our SAP R/3 data extraction requirement.  And this technical understanding is not clear to us as regards the difference between the 2 flows.  And being new to this whole of ETL - I just wanted to know the pros and cons of this particular data extraction. 
    As advised I shall check the schedules for a week, and then we shall move it probably into production.
    Thanks again.
    Kind Regards
    Indu
    Edited by: Indumathy Narayanan on Aug 22, 2011 7:02 PM

  • BODS 3.1 : How to trigger an email alert for the jobs on BODS server ?

    Hi All.
    I have this request.
    BODS 3.1 : How to trigger an email alert for the jobs on BODS server ?
    We have jobs scheduled on BODS running smoothly and absolutely fine.
    But to check, i am logging into the admin console and check for the jobs status.
    I would like to have an email to be received from BODS after each job is finished.
    It could succuessful. Or it could fail.
    Whatsoever, i wish to receive an email alert as soon as a job is finished.
    Can anyone advise me as to whether this could be made possible.
    And if yes, how this could be done.
    Thanks for your help in advance.
    In BOE CMC / for webi / schedule / we find an option to send email for a job success or a job failure.
    Is there any option similar to that in BODS ?
    Also would like to know :
    how to use the smtp_to or mail_to functions ?
    how to set up the smtp server for this ?
    thanks
    REgards
    indu
    Edited by: Indumathy Narayanan on May 31, 2011 3:47 PM

    Hi.
    Since am new to this BODS. I need some help.
    I already have many jobs which are running absolutely fine.
    And when a job runs, and finishes, am able to see the trace saying
    e.g. :
    Job_abc is completed successfully.
    We got the smtp service activated for our test server.
    and we hae a group email id.
    I have put the details of the smtp server / ip address / and said apply restarted.
    The i created a simple test script as below :
    print (' Before email ' );
    smtp_to('abc@company_name.com', 'Job ' || job_name() ||' on ' || host_name() || ' has FAILED',
    ' the job has failed', 0, 0);
    print('After Email ');
    It does send a email to as per smtp_to whatever email is specified.
    But how to differentiate between a job success
    And a job which has failed.
    I wish to have a mail which says on the subject :
    'Job ' || job_name() ||' on ' || host_name() || ' has completed successfully'
    ==> IF it is a success
    OR
    'Job ' || job_name() ||' on ' || host_name() || ' has failed'
    ==> if it has failed
    How to make the system identify, whether
    to send a success message or a error message whatever
    Could anyone advise.
    thanks
    indu

  • Sometimes my email headers get crossed with bodies of other emails (next in the list), so how do I stop this from happening?

    We use Thunderbird with the email header list (Inbox, etc) in an upper pane, and the body of the highlighted email in a separate lower viewer pane.
    We compact files whenever asked by the software.
    In a long list of emails, I click on one header (at the top of the list) to view it in the pane below. Once in a while, the headers get mixed up and do not preview the proper email in the viewer pane.
    I will highlight and view an email, and then delete it. When this glitch occurs, the body of the deleted email appears in the lower pane when I click on the next email header in the upper pane list. The viewer pane should be showing me the new body of the just-highlighted email, but instead it shows the deleted body of the previously-viewed email. Once that new header is clicked, and the wrong body appears in the viewer pane, the proper email body disappears forever. I have to call my customer to have them resend.
    If I click on other emails further down the list, the same body (of the deleted email) continues to show up each time, and all the correct bodies of all the clicked headers are gone. So, phone calls have to happen to ask all the affected emails to be resent. (So, we have learned to stop as soon as we see it happen the first time.)
    The only way we have found to fix this is to close Thunderbird, and then reboot the computer. When Thunderbird is re-opened, the problem goes away. But any lost email bodies are permanently gone.
    We can't seem to find any strange set of events that make this happen. It appears random, and varied. Sometimes every week, sometimes not for months.
    This happens on 2 computers. One is a 2 year old PC (Compaq), and one is a 3 year old HP Probook laptop. Both currently using Thunderbird 24.5.0. Both run Windows 7. Both have lots of memory. Both are very stable machines with very few problems. Seems to be more prevalent on the Compaq desktop than the HP Probook.
    Please help.
    Thanks in advance.

    Do you organise your email into separate folders, or do you just let it accumulate in the Inbox?
    I would strongly advise you not to let it sit in the Inbox.
    When you see this happening, try right-click on the affected folder, select '''Properties''' then '''Repair Folder'''.
    Large folders (actually, a folder in Thunderbird is stored as a file, so it is sensitive to file size limits set by the underlying file system) can be problematic. This is one reason not to let messages collect in one folder. I use Thunderbird's Archive folders so that accumulated mails over several years can happily coexist because they are effectively stored in many small folders. No single mail store folder then exceeds the 2GB or 4GB sizes that have been known to stress the OS. I appreciate that current builds of Thunderbird and a modern 64-bit OS should be able to cope, but practically I find it slows down when given huge files, so I err towards a pragmatic solution; a large number of not very big files.
    Archives are searchable, and a Saved Search folder can give you a virtual composite folder allowing access to the entire Archive.

  • MS Access and BODS

    Dear All,
    For one of my object, I need to read data from MS Access tables using SAP BODS tool. At a particular location, I have MS Access tables stored and I need to go to that path and extract the data present in these tables and load them to target table in SAP BODS.
    As per my understanding, I need to create a Datastore which can read data from these tables. Could anyone please help me in creating that Datastore or let me know the process/procedure how to create one?
    Also what all details I would need in order to create a Datastore which can access these MS access tables?
    Thanks,
    Surbhi Singi

    Hi Surbhi,
    Follow these steps:
    Firstly check the configurion of your MS Access ODBC connection.
    You can do this by going to
    control panel->administrative tools->data sources(ODBC)->SYSTEM DSN
    Check for MS ACCESS DSN. If not create one.
    Secondly open your Data Services Designer -> Datastore tab(below project explorer) -> Rightclick->New Data store
    Enter data store name, Select Data store Type as Database and Database type as ODBC.
    In the Data source name select the MS ACESS DSN created in the first step.
    Enter the credentials and click ok.
    You are done!!
    Later you can import the respective tables and load the data.
    Regards,
    Mubashir Hussain

  • Using the Execute Preloaded Option for ABAP Dataflows in SAP BODS

    Hello All,
    This is regarding the use of Advanced Option in SAP Application Type (ECC) Datastore settings in SAP BODS 4.2, where there are two options to select from for ABAP Execution Option property: (1)Generate and Execute and (2) Execute Preloaded. Since our ECC client is often locked by BASIS team even on DEV environment, we would like to make use of the second option 'Execute Preloaded' so that we could extract the data from ECC tables without having to ask the BASIS team to unlock the ECC client every time before extraction.
    The problem is that we are getting an error upon generating and uploading the ABAP program to ECC client. I have searched the blogs and so far I have only found that there are certain ABAP programs or function modules that come with SAP BODS which need to be installed by BASIS team on ECC side to allow the ABAP dataflows to be generated and uploaded to ECC server. I would appreciate it if anyone could provide a list of which function modules BASIS needs to install on ECC or a blog that provides details around using this option.
    So far, from the SAP BODS designer, we are performing below steps but getting an error upon generating and uploading the ABAP dataflow program:
    1. Create a test ABAP Dataflow using SAP ECC datastore. Provide the ABAP program options.
    2. Right click, select Generate ABAP Program.
    3. Once the ABAP Program Generation Dialog box appears, check the box "Upload Generated Program".
    4. Upon clicking OK, we are getting the following error:
    The ABAP program <ZRTEST01> for ABAP data flow <RT_TEST_R3> (datastore <R3_DS>) was not uploaded: < RFC CallReceive error <Function /BODS/RFC_ABAP_INSTALL_AND_RUN: RFC_ABAP_MESSAGE- Exception condition "NOT_SUPPORTED_BY_GUI" trigger[SAP NWRFC 720][SAP Partner ### ][clientname][servername][accountname][4103]>. >.
    Any help would be greatly appreciated.
    Thanks,
    Rizwan

    All,
    The BASIS team reviewed steps provided in the BODS document and attempted to install the Function Modules but now none of the BODS jobs would work. All BODS jobs are being terminted with the Syntax Error error when extracting data from ECC using ABAP dataflows:
    17740 16404 R3C-150412 06/17/2014 9:38:46 AM |Data flow RT_DF_TEST_PC207
    17740 16404 R3C-150412 06/17/2014 9:38:46 AM RFC CallReceive error <Function /BODS/RFC_ABAP_INSTALL_AND_RUN: RFC_ABAP_RUNTIME_FAILURE-(Exception_Key: SYNTAX_ERROR)- Syntax
    17740 16404 R3C-150412 06/17/2014 9:38:46 AM error in program /BODS/SAPLBODS                          .[SAP NWRFC 720][SAP Partner 740 ][DEV1][Server][account][4103]>.
    22052 20572 R3C-150412 06/17/2014 9:38:46 AM |Data flow RT_DF_TEST_PC207
    22052 20572 R3C-150412 06/17/2014 9:38:46 AM RFC CallReceive error <Function /BODS/RFC_ABAP_INSTALL_AND_RUN: RFC_ABAP_RUNTIME_FAILURE-(Exception_Key: SYNTAX_ERROR)- Syntax
    22052 20572 R3C-150412 06/17/2014 9:38:46 AM error in program /BODS/SAPLBODS                          .[SAP NWRFC 720][SAP Partner 740 ][DEV1][Server][account][4103]>.

  • BODS 3.1 :SMTP configuration to send email for success or failure of job

    Hi.
    This is further to this post : http://forums.sdn.sap.com/thread.jspa?threadID=1965129&tstart=0
    Thanks to all for the help.
    I did not find a complete set of information as things to do (for a person who is new to BODS).
    Hence sharing the workflow which works for me when tested on a simple workflow  :
    Pre-requisite : SMTP service to be made available on the BODS server and port 25 to be opened up.
    This is done by the IT dept of the company.
    When smtp service is activated and port 25 is opened up, telnet IP address 25 should say, service is available.
    Suggestion : create a email group like : bods at company.com which can be used to be used for the BODS JOBs. Which does not prompt for a password. This also is done by the IT dept of the company.
    1) On the BODS server :
    1) Click open Data Services Server Manager
    2) SMTP settings for smtp_to functions :
    Add the IP Address of the SMTP server
    Add the email address of the sender. Eg. bodsjobs at companyname,com
    3)Apply/Restart the server.
    2) Login to BODS Designer :
    Create a test job :
    TRY ==>DF===> Script for success ===> CATCH script for failure
    In the script for success : add the following :
    print (' IN TRY BLOCK Before email ' );
    smtp_to('bodsjobs at abc,com', ' SUCCESS OF JOB  :  '
    || job_name(),'Job completed successfully ', 0, 0);
    print('After Email ');
    In the script for Catch add the following :
    print (' IN CATCH BLOCK Before email ' );
    smtp_to('bodsjobs at abc,com', 'FAILURE OF JOB : '
    || job_name(),
    error_number( ) || 'Error message: ' || error_message( ) ||
    'check details', 50, 50);
    print('After Email ');
    Note : In the catch : select all exceptions on the check box.
    Save all.
    Click on the job and say execute.
    The job will run and if it is successful. You will receive the succes mail.
    If there is any error in the DF before the catch, you will receive a mail, which says failure.
    The above workflow, when tested atleast works for me.
    Hope this  helps.
    Regards
    indu

    Hi Arun.
    Am really not aware as to how to make things work from the console.
    I work on windows environment both at client and on the server.
    It looks like you got a bit confused on suggestion of a email group account.
    I never did anything on outlook configuration for activation of smtp for bods.
    I think you should not mess with outlook.
    I dont know from the client side how things work.
    But i have worked directly on the BODS server where the bods installation was done.
    It is relatively very simple to do and to test.
    But to even test why you get that error.
    All you need is just a email id. Which could be yours ???
    Put your email id as the source of email in that box in the job server.
    And put only your email id in smtp_to and test it.
    Run a test job. If it completes successfully. You will receive a mail.
    In the catch, if you have checked all the boxes for error capture,
    if the job fails, it would throw a dump of the errors in the email.
    Once everything works well and you receive emails for any jobs you are running.
    Then you could decide for the source of your email id and for the receipients of your emails of the jobs.
    You could put a group id of your team as source. And also in smtp_to script in case you want everyone to receive the alert that the job has been done or it has errors.
    Or include only those email ids in the smtp_to whom you feel, should receive the mail.
    Hope this idea helps.
    Please dont mess with outlook for any smtp activation for bods.
    It is not required if you are working on a windows environment.
    You just need a email id for testing.
    Outlook configuration does not come into picture at all.
    if there is any other unix environment involved. I have no idea.
    And i really do not know how things would be from a client side to configure.
    Server configurations i do from the server.
    I did all configurations on the main server machine. Not from the client.
    And I have only one job server so i did everything there.
    I did not do anything from management console.
    I use management console only for monitoring of jobs on a daily basis, the dashboard which gives you details of the data pulled, time taken etc etc.
    cheers
    indu.
    Edited by: Indumathy Narayanan on Oct 22, 2011 8:05 AM
    Edited by: Indumathy Narayanan on Oct 22, 2011 8:09 AM
    Edited by: Indumathy Narayanan on Oct 22, 2011 8:17 AM
    Edited by: Indumathy Narayanan on Oct 22, 2011 8:18 AM

  • How to load data into 3 different target tables usin BODS ?

    Hello Friends,
    I have 5 different source tables with same field definitions, Now I want to load all the records into three/four different target tables (Flat file, SQL Server, XML, and Oracle ) Could anyone please tell me how to do this task ?
    Thanks in Advance,
    Bheem.

    Hello Bheem,
    You can create separated dataflow for each target as suggested by Bala, this is a good choice when evaluating you scenario.
    If you put all targets in the same dataflow you may experience problems if one of them is down (as you have different servers as targets).
    BODS will send the data simultaneously to all targets and when one fails, the load will be break and you may have the tables not sinc anyways (if you plan to use a single dataflow to avoid this situation, it won't do it).
    So the best option is to create separated dataflow and put inside another one so you can run a singe dataflow that will call each one of the target and if onw fails the other will complete accordingly.
    Then if you put them inside a error trapping, you may even make your dataflow to retry the load prior abend the job.
    Think about cascading your dataflows and let the leaf level simple as it can be so it will be easier to schedule and debug your process.
    Pay attention to datatypes and other conversions you might need when working with more than one source/target.
    Regards,

  • How to email data in a table using BODI

    Dear Pals,
    I have a table with 1000 records, I want to send this data through mail as an attachment(EXCEL)  every day at 9:00 AM . Can we do it using BODI?
    Please do post if there is any way?
    many many thanks in advance.
    Regards,
    Suneel Dirasanchu

    Assuming BOE is consuming the data you have in your tables.
    - Create a WebI Report on this table via universe OR Create a Crystal Report (via Universe or directly ) on the table and save it to BOE.
    - Create a file event in BOE for this webi report to be kicked off.
    - In the DI Job, use the exec command to place the file in the location that the file wait event on BOE is looking for.
    - If you place the WebI or Crystal report to be scheduled to wait for the file event and choose the type of format (Excel, webI, etc) and destination(email as attachment).
    I implemented this as part of hand off from DI to BOE system for a recent project. The WebI reports have been created by the reporting team and I place the files via the exec command. The reports are scheduled to run and send the email as an attachment.
    Hope this helps!
    Thanks & Regards
    Tiji

  • How to create different bodies in a single region - Urgent

    A simple Question !
    How to create more than 1 body in a single region. E.g i created report region and under body tag in region all other charts reports are created.
    i dont want that .. i want to create different bodies in a single region.
    Guide pls

    See this link:
    http://apex.oracle.com/pls/otn/f?p=31517:107
    Denes Kubicek
    http://deneskubicek.blogspot.com/
    http://www.apress.com/9781430235125
    http://apex.oracle.com/pls/otn/f?p=31517:1
    http://www.amazon.de/Oracle-APEX-XE-Praxis/dp/3826655494
    -------------------------------------------------------------------

  • How to get bode plot in lab view

    Hi all,
    I'm just starting with lab view.
    We have NI PXI system.
    I need frequency responce of a servo motor i.e.  bode plot,
    I'm using labview 8 ,we have PXI 7352 motion controller.
    I'm very confused as where to start and what is the in put to the motor.
    can any one help me?

    Hello,
    I'm not so sure that using a motion control card is the way to go about finding the frequency response of a servo motor. In order to do this, you would give the motor a varying sinusoidal voltage and then measure the voltage output of a velocity sensor. Here is a link to a document that goes through the process: ftp://ftp.ni.com/pub/devzone/epd/lab5_frequency_response.pdf
    -Zach
    Certified LabVIEW Developer

  • Unknown error calling a RFC Function Module in BODI

    Hi all,
    I am having an issue identifying the error that my batch job in BODI is giving with regards to a RFC function module call in my query. The value returned by the standard parameter 'AL_RFC_RETCODE' are square characters (              ).
    At the completion of the job, the error message 'RFC CallReceive error <Function ZCT_INDSLS_SLS_TRN_INSERT: Invalid request.. Invalid request.>.' is being output.
    Correct me if I'm wrong but the parameter 'AL_RFC_RETCODE' should return standard values as per the techical manual provided to undertand any errors that might occur during a RFC FM call.
    Can anyone help me on this?
    Best regards,
    Andrew Wong

    Hi Senthil,
    I think you answered your query in your question itself
    The user ID and password need to be maintained in RFC destination in SM59.
    Go to SM59 and then click on test connection..if it does not ask for ID and passwoed that means correct ID and password is maintained in SM59 otherwise maintain it.
    Regards,
    Atish

  • Get a connection error when running a BODS job

    Hi All,
    I have BODS system setup in my environment. There are quite some jobs that run everyday. The jobs have stopped running since the last days. The jobs arn't failing, but they do not run.
    When I try to Execute a job, I get the error "Connection Refused: connection". When I Test the connection from BODS to CMS, it is working fine. Also, from the BODS/BO server I am able to connect/ping to the DataStore servers. My Datastores are SQL servers and ECC also.
    I do not know what is the resolution for the issue.
    Thanks,
    Amrita

    Hi Manoj,
    We recognized the issue was at the OS level, in the "Data Services"-OS Service.
    The BODS jobs are running fine now.
    Thanks,
    Amrita

  • Error while using Extractors in BODS

    Hi All,
    1.Our version is BODS 4.0  & ECC 6.0 EHP5.
    2. We can able to import extractor from SAP system.
    3.Error while running the extractor job in BODS, Please see the error below.
    Extractor returned error: <E: MESSAGE - Job ODQR_20140806_121712_000014_C could not be scheduled. See the system log for
                                                           details.. MESSAGE_V1 - ODQR_20140806_121712_000014_C.>.
    4. I checked running RSA3 transaction, there we can able to see the data. Aslo tested by using ABAP program, the result was Maximum waiting time of 7.200 seconds exceeded.
    Aslo i refered below link.
    http://wiki.scn.sap.com/wiki/pages/viewpage.action?pageId=272794989
    Thanks & Regards,
    Balamurugan G

    Hi Ganesh,
    Its standard extractor - 0CUSTOMER_ATTR
    I can able to see the data in RSA3 but i don't know how to check in RSA5/RSA6 whether its enabled or not.
    I checked in ROOSATTR table that extractor is available.
    Thanks & Regards,
    Balamuruga G

Maybe you are looking for

  • Import itab in memory in user exit

    I am trying to export an internal table to memory in BADI and import from memory into an internal table in User exit . It looks like the export works, but when I import in program2, the internal table is empty. Even when the command returns sy-subrc

  • Material Ledger config

    Hi, I need to activated material ledger with different exchange rates. I never done any material ledger config. Can anyone help me regarding the configuration steps from material ledger with different exchange rates. Any input would be appreciated. T

  • Unlock iphone4 by officialiphoneunlock.uk ?.

    Quelqu un connais ce site etsavoir si cserieux car j attent depuis le11/12/13 et toujours rien

  • Package's name change? / makepkg can't fetch sources from my URL!

    My next package to submit to the AUR has had an official namechange from z80 Assembly IDE to Wabbitcode. The program still displays the old name; however, it is going to be changed soon. Therefore I am torn on what name to use, seeing as it does not

  • Date Diff problem

    Hi, I'm writing a java app which is connected to a ms access db. I have the following query, but i get a run time error that states "too few parameters expected 1". I think the problem lies with the DateDiff part of the query, i want it to calculate