How to see data in BAM from sensor on BPEL activity?

Hello, I 'am working with BPEL on Jdeveloper (last version).
I have putted sensors on some activity of my process. On an other machine I have putted the BAM server (last version) which is ON and I have the BAM connection on Jdevlepper. Now I try to see data on BAM from sensor but I can't.
Can someone explain me how to do?
(I just can build some data just like in a data base with BAM architect but I can't read and put to graphics the ones from the sensors)
Thank's much
JAck

Hello,
To answer your question, I have install BAM on the same machine where BPEL server is running.
Now I steel can't get sensor's data from BPEL:
With BAM architect I can build data, with BAM studio I can use that data to design graphics but when I try to feed data from BPEL it doesn't work because I just can't press the OK button!!
In BPEL : When I try to Create BAM sensor action (from structure panel), I can't generate the xsl map file no action is done when pressing OK, create mapping or edit mapping button!

Similar Messages

  • How to view data in Xcelsius from an external Excel file

    Hi Everyone,
    i wonder if you know How to view data in Xcelsius from external Excel file and not depend only on the emebded excel sheet.

    Hi,
    As per my understanding, excel has to be imported in Xcelsius to draw up the dashboard. Your aggregated data sits in the xlf file. Using an external excel file will not be possible (I guess).
    One of the ways to connect to an external excel, is the use XML Maps but then your data has to be in the form of XML which will be pulled into excel first and then needs to be brought in Xcelsius using Data Manager.
    You can explore XML Maps in xcelsius in SDN.
    Regards
    Nikhil

  • How to send data to bam data object through java code

    how to send data to bam data object through java code

    I've made a suggestion in other thread: https://forums.oracle.com/thread/2560276
    You can invoke BAM Webservices (Using Oracle BAM Web Services) or use JMS integration using Enterprise Message Sources (http://docs.oracle.com/cd/E17904_01/integration.1111/e10224/bam_ent_msg_sources.htm)
    Regards
    Luis Fernando Heckler

  • How to download data in excel from web report in sap ?

    how to download data in excel from web report made  in sap abap?
    through tcode smw0.

    for exemple using
    MS_EXCEL_OLE_STANDARD_DAT

  • How-to extract data in chunks from one R3 to another R3 system

    Does anybody know how to extract data in chunks from one system to another? Imagine you would like to extract 1M records in chunks of 10K records, how would you do it between 2 different R3 systems? A cursor (even with hold) doesn't work
    Any suggestion appreciated
    Thanks
    Ioan

    Hi Ioan,
    I'd do it with fm : TABLE_ENTRIES_GET_VIA_RFC (with where clause-tab)
    or fm GET_TABLE_RFC (get all items)
    regards Andreas

  • How to see data modeller reports in sql developer?

    How to see data modeller reports in sql developer?

    1) export your design to existing schema in Oracle database
    2) define connection to that schema in SQL Developer
    3) find "Data Modeler Reports" in reports window of SQL Developer - probably "design rules" are more interesting here
    Philip

  • How to see work flow log from Work flow number & work item?

    Hi SDN,
    How to see work flow log from Work flow number & work item?
    Please let me know if there is any transaction for this?
    Or there is any table which gives the details of Work items & work flow numbers?
    Regards,
    Rahul
    Edited by: Rahul Wagh on Nov 12, 2008 5:56 PM

    Hi Rahul
    Go to the Transaction SWI1,
    Enter the workitem no and press F8.
    Select the required entry and Press Shift+F8 (Log Button)
    to view the workflow log.
    Best Regards,
    Deepa Kulkarni

  • How to Extract data and reports from the SAP BW warehouse

    Hi to All Guru's
    I am new to SAP BW. Can any one help me to know, How to Extract data and reports from SAP BW. Do we use any tools.  After Extracting the data from SAP BW how can we move all the data to Cognos.  It would be kind enough to provide me with any documentation or links.  Step by step procedure would be very much helpfull to me.
    Thanks
    Venu

    Hi Voodi
    Thanks for the quick response. I think Open hub is to extract the data from SAP BW, but what about sending this data into Cognos.  Can you please let me know in detail regarding answer. If possible send me any documentation or links regarding this.  Thanks for your concern.
    Venu

  • Cannot see data in BSO from transparent ASO partition

    I have 2 cubes: BSO (target) and ASO (source).
    BSO should show data from ASO corresponding element "Fact" from dimension "Scenario".
    I can see the data using ASO directly, but I cannot see data viewing it from BSO using transparent partition...

    Hi,
    When you have the data in the ASO( i.e source cube) ,it should reflect in the BSO ( i.e destination) . I am sure there is problem with the transparent partition ,you had created .Validate and share more info of error ( if you encounter during partition creatino) , that should be resourceful
    Sandeep Reddy Enti
    HCC

  • How to see data for particular date from a alert log file

    Hi Experts,
    I would like to know how can i see data for a particular date from alert_db.log in unix environment. I'm suing 0racle 9i in unix
    Right now i'm using tail -500 alert_db.log>alert.txt then view the whole thing. But is there any easier way to see for a partiicular date or time
    Thanks
    Shaan

    Hi Jaffar,
    Here i have to pass exactly date and time, is there any way to see records for let say Nov 23 2007. because when i used this
    tail -500 alert_sid.log | grep " Nov 23 2007" > alert_date.txt
    It's not working. Here is the sample log file
    Mon Nov 26 21:42:43 2007
    Thread 1 advanced to log sequence 138
    Current log# 3 seq# 138 mem# 0: /oracle/NEWDB/oradata/NEWDB/redo3.log
    Mon Nov 26 21:42:43 2007
    ARCH: Evaluating archive log 1 thread 1 sequence 137
    Mon Nov 26 21:42:43 2007
    ARC1: Evaluating archive log 1 thread 1 sequence 137
    ARC1: Unable to archive log 1 thread 1 sequence 137
    Log actively being archived by another process
    Mon Nov 26 21:42:43 2007
    ARCH: Beginning to archive log 1 thread 1 sequence 137
    Creating archive destination LOG_ARCHIVE_DEST_1: '/oracle/NEWDB/admin/arch/1_137
    .dbf'
    ARCH: Completed archiving log 1 thread 1 sequence 137
    Mon Nov 26 21:42:44 2007
    Thread 1 advanced to log sequence 139
    Current log# 2 seq# 139 mem# 0: /oracle/NEWDB/oradata/NEWDB/redo2.log
    Mon Nov 26 21:42:44 2007
    ARC0: Evaluating archive log 3 thread 1 sequence 138
    ARC0: Beginning to archive log 3 thread 1 sequence 138
    Creating archive destination LOG_ARCHIVE_DEST_1: '/oracle/NEWDB/admin/arch/1_138
    .dbf'
    Mon Nov 26 21:42:44 2007
    ARCH: Evaluating archive log 3 thread 1 sequence 138
    ARCH: Unable to archive log 3 thread 1 sequence 138
    Log actively being archived by another process
    Mon Nov 26 21:42:45 2007
    ARC0: Completed archiving log 3 thread 1 sequence 138
    Mon Nov 26 21:45:12 2007
    Starting control autobackup
    Mon Nov 26 21:45:56 2007
    Control autobackup written to SBT_TAPE device
    comment 'API Version 2.0,MMS Version 5.0.0.0',
    media 'WP0033'
    handle 'c-2861328927-20071126-01'
    Clearing standby activation ID 2873610446 (0xab47d0ce)
    The primary database controlfile was created using the
    'MAXLOGFILES 5' clause.
    The resulting standby controlfile will not have enough
    available logfile entries to support an adequate number
    of standby redo logfiles. Consider re-creating the
    primary controlfile using 'MAXLOGFILES 8' (or larger).
    Use the following SQL commands on the standby database to create
    standby redo logfiles that match the primary database:
    ALTER DATABASE ADD STANDBY LOGFILE 'srl1.f' SIZE 10485760;
    ALTER DATABASE ADD STANDBY LOGFILE 'srl2.f' SIZE 10485760;
    ALTER DATABASE ADD STANDBY LOGFILE 'srl3.f' SIZE 10485760;
    ALTER DATABASE ADD STANDBY LOGFILE 'srl4.f' SIZE 10485760;
    Tue Nov 27 21:23:50 2007
    Starting control autobackup
    Tue Nov 27 21:30:49 2007
    Control autobackup written to SBT_TAPE device
    comment 'API Version 2.0,MMS Version 5.0.0.0',
    media 'WP0280'
    handle 'c-2861328927-20071127-00'
    Tue Nov 27 21:30:57 2007
    ARC1: Evaluating archive log 2 thread 1 sequence 139
    ARC1: Beginning to archive log 2 thread 1 sequence 139
    Creating archive destination LOG_ARCHIVE_DEST_1: '/oracle/NEWDB/admin/arch/1_139
    .dbf'
    Tue Nov 27 21:30:57 2007
    Thread 1 advanced to log sequence 140
    Current log# 1 seq# 140 mem# 0: /oracle/NEWDB/oradata/NEWDB/redo1.log
    Tue Nov 27 21:30:57 2007
    ARCH: Evaluating archive log 2 thread 1 sequence 139
    ARCH: Unable to archive log 2 thread 1 sequence 139
    Log actively being archived by another process
    Tue Nov 27 21:30:58 2007
    ARC1: Completed archiving log 2 thread 1 sequence 139
    Tue Nov 27 21:30:58 2007
    Thread 1 advanced to log sequence 141
    Current log# 3 seq# 141 mem# 0: /oracle/NEWDB/oradata/NEWDB/redo3.log
    Tue Nov 27 21:30:58 2007
    ARCH: Evaluating archive log 1 thread 1 sequence 140
    ARCH: Beginning to archive log 1 thread 1 sequence 140
    Creating archive destination LOG_ARCHIVE_DEST_1: '/oracle/NEWDB/admin/arch/1_140
    .dbf'
    Tue Nov 27 21:30:58 2007
    ARC1: Evaluating archive log 1 thread 1 sequence 140
    ARC1: Unable to archive log 1 thread 1 sequence 140
    Log actively being archived by another process
    Tue Nov 27 21:30:58 2007
    ARCH: Completed archiving log 1 thread 1 sequence 140
    Tue Nov 27 21:33:16 2007
    Starting control autobackup
    Tue Nov 27 21:34:29 2007
    Control autobackup written to SBT_TAPE device
    comment 'API Version 2.0,MMS Version 5.0.0.0',
    media 'WP0205'
    handle 'c-2861328927-20071127-01'
    Clearing standby activation ID 2873610446 (0xab47d0ce)
    The primary database controlfile was created using the
    'MAXLOGFILES 5' clause.
    The resulting standby controlfile will not have enough
    available logfile entries to support an adequate number
    of standby redo logfiles. Consider re-creating the
    primary controlfile using 'MAXLOGFILES 8' (or larger).
    Use the following SQL commands on the standby database to create
    standby redo logfiles that match the primary database:
    ALTER DATABASE ADD STANDBY LOGFILE 'srl1.f' SIZE 10485760;
    ALTER DATABASE ADD STANDBY LOGFILE 'srl2.f' SIZE 10485760;
    ALTER DATABASE ADD STANDBY LOGFILE 'srl3.f' SIZE 10485760;
    ALTER DATABASE ADD STANDBY LOGFILE 'srl4.f' SIZE 10485760;
    Wed Nov 28 21:43:31 2007
    Starting control autobackup
    Wed Nov 28 21:43:59 2007
    Control autobackup written to SBT_TAPE device
    comment 'API Version 2.0,MMS Version 5.0.0.0',
    media 'WP0202'
    handle 'c-2861328927-20071128-00'
    Wed Nov 28 21:44:08 2007
    Thread 1 advanced to log sequence 142
    Current log# 2 seq# 142 mem# 0: /oracle/NEWDB/oradata/NEWDB/redo2.log
    Wed Nov 28 21:44:08 2007
    ARCH: Evaluating archive log 3 thread 1 sequence 141
    ARCH: Beginning to archive log 3 thread 1 sequence 141
    Wed Nov 28 21:44:08 2007
    ARC1: Evaluating archive log 3 thread 1 sequence 141
    ARC1: Unable to archive log 3 thread 1 sequence 141
    Log actively being archived by another process
    Wed Nov 28 21:44:08 2007
    Creating archive destination LOG_ARCHIVE_DEST_1: '/oracle/NEWDB/admin/arch/1_141
    .dbf'
    Wed Nov 28 21:44:08 2007
    ARC0: Evaluating archive log 3 thread 1 sequence 141
    ARC0: Unable to archive log 3 thread 1 sequence 141
    Log actively being archived by another process
    Wed Nov 28 21:44:08 2007
    ARCH: Completed archiving log 3 thread 1 sequence 141
    Wed Nov 28 21:44:09 2007
    Thread 1 advanced to log sequence 143
    Current log# 1 seq# 143 mem# 0: /oracle/NEWDB/oradata/NEWDB/redo1.log
    Wed Nov 28 21:44:09 2007
    ARCH: Evaluating archive log 2 thread 1 sequence 142
    ARCH: Beginning to archive log 2 thread 1 sequence 142
    Creating archive destination LOG_ARCHIVE_DEST_1: '/oracle/NEWDB/admin/arch/1_142
    .dbf'
    Wed Nov 28 21:44:09 2007
    ARC0: Evaluating archive log 2 thread 1 sequence 142
    ARC0: Unable to archive log 2 thread 1 sequence 142
    Log actively being archived by another process
    Wed Nov 28 21:44:09 2007
    ARCH: Completed archiving log 2 thread 1 sequence 142
    Wed Nov 28 21:44:36 2007
    Starting control autobackup
    Wed Nov 28 21:45:00 2007
    Control autobackup written to SBT_TAPE device
    comment 'API Version 2.0,MMS Version 5.0.0.0',
    media 'WP0202'
    handle 'c-2861328927-20071128-01'
    Clearing standby activation ID 2873610446 (0xab47d0ce)
    The primary database controlfile was created using the
    'MAXLOGFILES 5' clause.
    The resulting standby controlfile will not have enough
    available logfile entries to support an adequate number
    of standby redo logfiles. Consider re-creating the
    primary controlfile using 'MAXLOGFILES 8' (or larger).
    Use the following SQL commands on the standby database to create
    standby redo logfiles that match the primary database:
    ALTER DATABASE ADD STANDBY LOGFILE 'srl1.f' SIZE 10485760;
    ALTER DATABASE ADD STANDBY LOGFILE 'srl2.f' SIZE 10485760;
    ALTER DATABASE ADD STANDBY LOGFILE 'srl3.f' SIZE 10485760;
    ALTER DATABASE ADD STANDBY LOGFILE 'srl4.f' SIZE 10485760;
    Thu Nov 29 21:36:44 2007
    Starting control autobackup
    Thu Nov 29 21:42:53 2007
    Control autobackup written to SBT_TAPE device
    comment 'API Version 2.0,MMS Version 5.0.0.0',
    media 'WP0206'
    handle 'c-2861328927-20071129-00'
    Thu Nov 29 21:43:01 2007
    Thread 1 advanced to log sequence 144
    Current log# 3 seq# 144 mem# 0: /oracle/NEWDB/oradata/NEWDB/redo3.log
    Thu Nov 29 21:43:01 2007
    ARCH: Evaluating archive log 1 thread 1 sequence 143
    ARCH: Beginning to archive log 1 thread 1 sequence 143
    Creating archive destination LOG_ARCHIVE_DEST_1: '/oracle/NEWDB/admin/arch/1_143
    .dbf'
    Thu Nov 29 21:43:01 2007
    ARC1: Evaluating archive log 1 thread 1 sequence 143
    ARC1: Unable to archive log 1 thread 1 sequence 143
    Log actively being archived by another process
    Thu Nov 29 21:43:02 2007
    ARCH: Completed archiving log 1 thread 1 sequence 143
    Thu Nov 29 21:43:03 2007
    Thread 1 advanced to log sequence 145
    Current log# 2 seq# 145 mem# 0: /oracle/NEWDB/oradata/NEWDB/redo2.log
    Thu Nov 29 21:43:03 2007
    ARCH: Evaluating archive log 3 thread 1 sequence 144
    ARCH: Beginning to archive log 3 thread 1 sequence 144
    Creating archive destination LOG_ARCHIVE_DEST_1: '/oracle/NEWDB/admin/arch/1_144
    .dbf'
    Thu Nov 29 21:43:03 2007
    ARC0: Evaluating archive log 3 thread 1 sequence 144
    ARC0: Unable to archive log 3 thread 1 sequence 144
    Log actively being archived by another process
    Thu Nov 29 21:43:03 2007
    ARCH: Completed archiving log 3 thread 1 sequence 144
    Thu Nov 29 21:49:00 2007
    Starting control autobackup
    Thu Nov 29 21:50:14 2007
    Control autobackup written to SBT_TAPE device
    comment 'API Version 2.0,MMS Version 5.0.0.0',
    media 'WP0280'
    handle 'c-2861328927-20071129-01'
    Thanks
    Shaan

  • How to see data from anemometer in labview?

    I'm attempting to see the data from my anemometer via labview...when that didn't work I wanted to make sure the instrument was working, so I tried to set it up in HyperTerminal like the anemometer instruction manual suggested. Now HyperTerminal won't let me connect to the anemometer using that port (keeps saying another program is using Com5, and to try again later!)
    Any suggestions on how to just basically See the data in labview from my anemometer(or hyperterminal if you're better at trouble shooting that!!)??
    Thanks!

    It's possible that Com5 isn't the correct port to read the data from. Since the built in serial port on most PCs is setup to be Com1, you might just be talking to wrong port. Of course if you've changed the default system setup or have added some extra serial ports, this may not be the issue. Hyperterminal probably isn't working because you have opened the port in LabVIEW but have not properly closed it. The operating system won't let more than one application at a time to access a serial port. Try shutting down LabVIEW, then try Hyperterminal. That should release the port. You might need to reboot. If you still get the error in Hyperterminal, then you are trying to access a port that is already in use by something else.
    Since your anemometer is a serial instrument, the easiest way to simply read data from it would be to use the Instrument Assistant Express VI. (assumming you're using LabVIEW 7.0 or 7.1 here)The Instrument Assistant is pretty good at hepling you figure out how to talk to your device. But if you're going to be using this in an application, you should look into using the VISA functions.
    Have a look through the LabVIEW examples (look in the help menu for "Find Examples") for plenty of serial examples. And remember to always close the port at the end of your application and don't use the Abort button on the toolbar to stop your application. Put everything in a While loop and use a boolean Stop button. The examples will show how to set this up.
    Ed
    Ed Dickens - Certified LabVIEW Architect - DISTek Integration, Inc. - NI Certified Alliance Partner
    Using the Abort button to stop your VI is like using a tree to stop your car. It works, but there may be consequences.

  • How to see data in the multi-provider ?

    We have a multiprovider with the key figure as total stock, and its given as a non-cumulative kf which has a check mark next to it. Its getting data from the infocube to it, I mean there are 4 infocubes and 1 ods to the multi-provider. If I right click and see the infocube or the multi-provider and select the fields by rite click and I am unable to see the total stock kf. But if I click the actual multiprovider and click display instead of display data, I am able to see the keyfigures folder, if i expand that I see stock quantity folder and many other folders. If I expand stock quantity folder, there are 3 kf's ztotalstock, 0isstock and 0recstock. In those ztotalstok is non-cumulative checked. And I am unable to see the kf value when i rite click on the mp. So any suggestion to display the data by selecting a posting date in the mp. What my guess is , will there be any internal table/view to display the mp data. Please let me know any suggestions.
    Thanks for your time,

    Prasanth, entering the mp name in LISTCUBE is same as right click on the mp and click on display data, where it will give the selecting criteria for the fields. But here both will display the same fields, that is not my case. I am unable to see one of the key figure which is inside the mp, have you opened any mp. Please right click on it and choose display and then go to the key figures folder , if you open that you have lot of folders in it based on the category. IN my case its the stock kf folder and if you expand that there are three kf's out of the three, one is ztotal stock which is declared as non-cumulative kf, rite next to it, option is checked. So I think that is the reson we are not able to see ztotalstock kf , no matter in the listcube or display data option on the mp. Hope you got my situation. I know how to see the data in the multi-provider or infocube. But in this case its not displaying that ztotal stock keyfigure, but in the reporting we can see the data on the mp. My question is why am I not able to see the selection for that ztotal stock key figure , either in listcube or rite clik display data option. Did anybody face this situation.
    Thank You,

  • How to merge data in BW from R3 and a legacy source

    Hi
    I would like to combine data from R3 and from a legacy system in BW as follows: I have two similar info objects:
    One info object contains the stuff from R3 (which contains some history but basically the data is just really well maintained starting from Jan 1st 2008.
    The other info object contains data from a legacy HR data warehouse which contains data back to 2001 and up to Dec 31st 2007.
    I would like to combine in BW the two info objects in a multiprovider such that if I make a query before Jan 2008 then the data is taken from the 'legacy' infoobject and when I run a query as of Jan 2008 or later then the data should be taken from the 'R3' info object. (Note that if I specify a range e.g. Dec 2007 - Jan 2008 then data from both info objects are to be taken).
    A problem I see is that the DATEFROM of rows in the time dependent tables of the 'R3' info object is not as of Jan 1st 2008 but can be any date before. Thus I wonder how an overlap between the 2 sources can be avoided.
    Is there a way to acomplish this?
    Regards and thanks for your help, Felix

    The how to at following url shows how to store, retrieve and display characters of different languages from Oracle database onto a JSP page.
    http://otn.oracle.com/sample_code/tech/java/codesnippet/jdbc/nls/Globalization.html
    HTH
    Chandar

  • How to see data in a transaction ODS

    Hi all,
    I am populating data into my Transactional ODS through the function module RSDRI_ODSO_INSERT_RFC . But not able to see the data in it. as there is no manage option to the ODS. and by default an export datasource is created for the ODS. do i have to create one more standard ods and use this export datasource as update rules and see the data?
    Please let me know how to see the data. and also and detail document on Transaction ODS?.

    Hello Satish,
    Data loaded through Planning Application eg. In Transactional ODS will not have manage option or request. You can view the data using following methods.
    Since transactional ODS objects cannot be filled with BW data using staging (data is not supplied from the DataSources), they are not displayed in the Scheduler or in the Monitor. Transactional ODS objects can therefore not be updated in the same way as standard ODS objects.
    If you switch a standard ODS object that already has update rules available to a transactional one, the update rules are set as inactive and are no longer processable.
    As no change log is generated, no delta update of data stored at the end of the process is possible.
    You cannot set the indicator for BEx Reporting when creating a transactional ODS object.
    1.) In order to report this, you can create an InfoSet and then execute a BEx query for it.
    2) You can also download the data from your transactional ODS object using the download function. You can find this function in the administration of the ODS objects, Tab Page Contents -> Active Data.  Choose Execute to display the data. Using the main menu Edit -> Download, you can download the data in different formats.
    Hope it helps.
    San.

  • How to tell date of manufacture from serial number or how to look up?also when were 1q524 introduce

    how to tell date of manufacture of 1q524 NC701AA-ABA from serial number or how to look up?
    when were they first introduced?

    dlkegarise
    See here for the release date. For the serial number the typical format is to have the 4th, 5th, and 6th characters to reference when the product was built.  The first character is the year date (Y) and the next two (WW), the week in that year it was made.
    CCSYWWZZZZ  
    Hope this helps you and anwsers your questions.  Regards.
    I work for HP. When your problem has been solved, accept the solution by clicking the "Accept as Solution" button to help other members in the future!

Maybe you are looking for