Having multiple threads for receiving RTP streams

Hello,
Developing an audio conference server, I have come to think that if I manage to separate the different audioReceivers who receive the RTP streams the performance could improve.
At this moment I have the main program, so the main thread let�s say, who initializes a new audioRx object for each remote client.
Would separation of the different receivers into different threads improve the applications performance?
has anyone thought of this? or done something similar?
Thanks for your help.
bgl

i need help from about the RTP stream from the same port

Similar Messages

  • Get a event for a new received rtp stream in a player

    Hi!
    I'm trying to implement a RTP-player, that receives a AV-stream and play it. The special thing about this player should be, that even if the stream interrupts, the player wait's on the same IP and port for a new stream and open it in the SAME frame (not like jmstudio in new window).
    I try to catch the "ReceiveStreamEvent", so i can restart the player, but i don't get eny events for this. I tried to do it with a "RTPManager", but i don't know how.
    Does anybody has a example how to get the "ReceiveStreamEvent", so i know, the RTP-stream has been interrupted?
    Thanks
    Adam

    See AVReceive2 in JMF Solutions, in JMF web site

  • Having multiple threads in ios 8.1 with the same people in a group message.

    When I send an imessage out to a group, some of the group responses will come back through a different message with all the same people. This just means I have 3-4 different threads in my messages with all of the same people. It's also difficult to read if I get multiple and check them at a later time because I have to see when they came through to piece the conversation together. Anyone else?

    Hello mollyh7,
    I would start with checking the your settings in Settings > Messages and as long as Group Messages is turned on you are good. But also keep in mind that if anyone else in the group does not have this enabled then it does throw everything off. Take a look at the article below for more information. 
    Send a group message with your iPhone, iPad, or iPod touch
    http://support.apple.com/kb/HT5760
    Regards,
    -Norm G. 

  • Having multiple paths for REPORTS_PATH

    I would like to have 2 directories/paths for REPORTS_PATH (one for production & other for test). My Environment: oracle 10g AS on Solaris.
    1)vi reports.sh (Commented below lines)
    NLS_LANG=ARABIC_KUWAIT.AR8MSWIN1256; export NLS_LANG
    REPORTS_PATH=/appl/ora10gas/liveapp10g:$ORACLE_HOME/reports/templates:$ORACLE_HOME/reports/printers:$ORACLE_HOME/reports/winfont
    2)
    vi repas1.conf
    <environment id="LIVE">
    <envVariable name="NLS_LANG" value="ARABIC_KUWAIT.AR8MSWIN1256"/>
    <envVariable name="REPORTS_PATH" value="/appl/ora10gas/liveapp10g:/appl/ora10gas/ASHome_1/reports/templates:/appl/ora10gas/ASHome_1/reports/printers:/appl/ora10gas/ASHome_1/reports/winfont"/>
    </environment>
    <environment id="TEST">
    <envVariable name="NLS_LANG" value="ARABIC_KUWAIT.AR8MSWIN1256"/>
    <envVariable name="REPORTS_PATH" value="/appl/ora10gas/testapp10g:/appl/ora10gas/ASHome_1/reports/templates:/appl/ora10gas/ASHome_1/reports/printers:/appl/ora10gas/ASHome_1/reports/winfont"/>
    </environment>
    <engine id="rwEng" class="oracle.reports.engine.EngineImpl" initEngine="2" maxEngine="8" minEngine="2" engLife="50" maxIdle="30" callbackTi
    meOut="90000" defaultEnvId="LIVE">
    (3) vi formsweb.cfg
    [appltest]
    envFile=appltest.env
    $cp default.env appltest.env
    $vi appltest.env
    REPORTS_ENVID=TEST
    Still it cannot read from multiple locations. Please can any one guide me on this ?
    Thanks
    Edited by: userR12 on Apr 14, 2012 11:13 AM

    Not sure if this helps at all. I just went through a nightmare trying to get 11g forms/reports to work through the forms builder.
    What I discovered with that is that I indeed had to set the REPORTS_PATH in rwserver.conf... however there were 3 different files that I ended up having to put it in (not sure if had to be in all 3). And perhaps more importantly, there seems to be an order to how you put these xml statements in the rwserver.conf files.
    In my case, it wouldn't work at all until I put it right after the last "<engine id=" statement.
    Mine looks something like this (not that this is a snippet of the file in Windows, for running via Forms/Reports Builder):
    <engine id="rwURLEng" class="oracle.reports.urlengine.URLEngineImpl" maxEngine="1" minEngine="0" engLife="50"/>
    <environment id="default">
         <envVariable value="C:\Oracle\WLS11gR1U4\user_projects\domains\AISworkstation\" name="REPORTS_PATH"/>
         <envVariable value="AMERICAN_AMERICA.WE8ISO8859P1" name="NLS_LANG"/>
         <envVariable value="RRRR-MM-DD" name="NLS_DATE_FORMAT"/>
    </environment>
    <environment id="windev_xxxxxxxx">
    <envVariable name="REPORTS_PATH"
    value="M:\ADS\Dvlprs\xxxxxxx\11g\FormsApps\stage;M:\ADS\Dvlprs\xxxxxxx\11g\FormsApps"/>
    </environment>
    I'm still figuring this all out too (although I did get it working how I need.. just not sure if it is the optimal solution), so I'm not sure how much this will help you. But for sure I can say that there is an order to these XML tags that I'm not so sure was the same (or maybe not enforced) in 10g.
    Good luck

  • Pros & Cons for having multiple AS for the same VRF

    Hi,
    I just wonder if there is any added value of having mutiple private AS numbers for a single VRF, compared to the usual practise of having one single AS for the entire enterprise network of a give customer. The solution provider is setting a requirement for having 3 AS numbers to connect to the MPLS IP-VPN, one AS for Primary Data Center, and one for Disaster Recovery Center and one for all the 1,000+ branaches.
    Appreciate your input.
    Sam

    Hi,
    I don't see any benefit for using three different AS number. Only benefit is we need not to use as-override (or allowas-in)
    if having three different AS.
    I worked for on Provider who was not much strong in processes and was not having automated tool to configure customer links, i mean everything was manual. I see many times that though customer requested to use as-ovveride at provider link
    people generally forget to configure the same and at time of testing the usual fault was missing this command at provider end.
    But if your provider is standard ( in terms of processes) then You can use only one AS number. This is what i experienced. technically i dont see much benefit in having three different AS.
    Regards # Mahesh

  • Multiple Threads writing to the same stream and "Write End Dead"

    Hi, I'm wondering what is the proper method for having multiple threads write to the same stream.
    Right now I have PipedWriter that I give to several Threads.
    However the JLS specifies that when the thread writing to PipedWriter dies, an exception will be generated. Does it mean that I may get a "write end dead" exception when any of the threads dies?
    More specifically I'm trying to extract keywords from a set of pages and print out keywords with their contexts (using threads to avoid a single blocked pages from blocking entire process)

    Your topic says "stream" but your post says "writer" - there's a difference. Readers/writers use the native encoding (which is correct if, for example, you're reading .DOC files created in Windows) while streams use UTF-8 encoding (which is correct if you've got one java app talking to another).
    Multiple threads can output to the same source. You've probably seen really annoying examples as your own System.out.println()s get intermixed with other JVM output in the JVM's window. Synchronization is going to be a serious issue.
    As to the exception, why not just catch it and ignore it?

  • How to synchronize audio and video rtp-streams

    Hi everyone!
    I've tried to make one of the processors, processor1, control the other processor, processor2, with processor1.addController(processor2) but I get a
    javax.media.IncompatibleTimeBaseException. I need to synchronize audio and video rtp-streams because they are sometimes completely out of sync.
    In JMF API Guide I've read:
    "2. Determine which Player object's time base is going to be used to drive
    the other Player objects and set the time base for the synchronized
    Player objects. Not all Player objects can assume a new time base.
    For example, if one of the Player objects you want to synchronize has
    a *push-data-source*, that Player object's time base must be used to
    drive the other Player objects."
    I'm using a custom AVReceive3 to receive rtp-streams and then I create processors for the incoming stream's datasource's, and they are ALL PushBufferDataSource's.
    Does this mean I can't synchronize these. If so is there any way I can change them into Pull...DataSources ?
    How are you supposed to sync audio and video if not with as above ?

    camelstrike wrote:
    Hi everyone!
    I've tried to make one of the processors, processor1, control the other processor, processor2, with processor1.addController(processor2) but I get a
    javax.media.IncompatibleTimeBaseException. I need to synchronize audio and video rtp-streams because they are sometimes completely out of sync.
    In JMF API Guide I've read:
    "2. Determine which Player object's time base is going to be used to drive
    the other Player objects and set the time base for the synchronized
    Player objects. Not all Player objects can assume a new time base.
    For example, if one of the Player objects you want to synchronize has
    a *push-data-source*, that Player object's time base must be used to
    drive the other Player objects."
    I'm using a custom AVReceive3 to receive rtp-streams and then I create processors for the incoming stream's datasource's, and they are ALL PushBufferDataSource's.
    Does this mean I can't synchronize these. If so is there any way I can change them into Pull...DataSources ?The RTP packets are timestamped when they leave, and they are played in order. You can't change the timebase on an RTP stream because, for all intensive purposes, it's "live" data. It plays at the speed the transmitting RTP server wants it to play, and it never gets behind because it drops out of order and old data packets.
    If your RTP streams aren't synced correctly on the receiving end, it means they aren't being synced on the transmitting end...so you should look into the other side of the equation.

  • If ScheduledThreadPoolExecutor provides multiple threads

    Hello,
    As per my understanding ScheduledThreadPoolExecutor, provides multiple threads for executing deferred and periodic tasks.
    public class TimerSample {
         public static void main(String args[]){
              ScheduledThreadPoolExecutor timerNew = new ScheduledThreadPoolExecutor(10);
              timerNew.scheduleAtFixedRate(new TimerTaskSample(), 0, 10 * 1000, TimeUnit.MILLISECONDS);
    class TimerTaskSample extends TimerTask{
         public void run(){
              System.out.println(Calendar.getInstance().getTimeInMillis());
              System.out.println(Thread.currentThread().getName());
              try {
                   Thread.sleep(60*1000);
              } catch (Exception e) {
                   e.printStackTrace();
    If this is understanding is correct then I should get a System,out statement after every 10 seconds. But I am getting sys out statement after every 60 seconds.
    Could you please explain this behavior

    The multiple threads is for running multiple tasks. It doesn't make a task multi-threaded. You are right that there is no point having more threads than tasks.
    For what you are trying to do you can generate tasks.
    final ScheduledThreadPoolExecutor timerNew = new ScheduledThreadPoolExecutor(10);
    timerNew.scheduleAtFixedRate(new Runnable() {
      public void run() {
        timerNew.submit(new TimerTaskSample());
      }}, 0, 10 * 1000, TimeUnit.MILLISECONDS);

  • Rtp stream applet

    Hi !!
    I'm new in JMF. I'd like to play rtp stream on remote machine by using applet . I've searched a lot for making java applet which could receive rtp stream but coldn't find easy to understand tutorial code etc.
    If you have any good links, please write it on forum or paste code.
    Thank you for any answers !!!

    an applet needs nothing great ...As someone who ..
    - has deployed 1.1-1.4 compatible applets,
    applets that interact with JavaScript, full (browser)
    window applets, applets that load screensavers
    and launch applications, web-started applets,
    applets that change PLAF..
    - has dealt with (literally) hundreds of applet
    related problems on usenet, including the
    investigation of applet bugs, and (particularly)
    bugs caused by the MSVM.
    ..I can assure you.
    1) Applets are anything but easy to deploy
    successfully on the wilds of the internet.
    2) As someone who has dealt a little with the
    JMF, I can add that JMF applets add a whole
    new level of complications to the mix.
    Generally, the end user needs to have
    pre-installed JMF in order to see JMF
    based applets. Even then, the browser
    might not detect the JMF classes.
    It is generally far more sensible, and reliable,
    to launch such projects in a free floating frame
    using web-start.
    Here is an example of a webstart launch.
    <http://www.javasaver.com/testjs/jmf/#test3>
    Note that any such web-start launch might be
    optimised in a variety of ways (to download less).
    The most obvious would be to use the
    customizer.jar to make a cut-down version
    that handles only the required sources (e.g. RTP)
    and formats (e.g. MOV).

  • Can I combine multiple threads from the same contact in iMessage?

    iMessage displays multiple threads for the same contact.  Its very confusing and has resulted in lost messages.  I've tried changing a number of settings but to no avail.  Seems to require everyone in my contact list to eliminate all their email addresses in iMessage.
    Begs the question why does iMessage even need this information?  Worse, why does it suck it into its settings by default.  The directory server should only require a valid Apple ID (associated with the users' devices)  in order to direct an iMessage between two contacts.

    Thanks for the reply, appreciate your input.  I did find another way of doing it and that is to highlight the first e-mail, hold down the shift key and click on the last one.  That highlights them all and then they can all be deleted.  Your ideas also work of course and I thank you for that.  Cheers.

  • Display a RTP Stream in JavaFX

    Hi all,
    does anyone know a way to display a RTP Stream? I´ve read about, that you can display a RTSP Stream as media.source.
    Is there something similar for a RTP Stream?
    Thanks in Advance
    Jazzor

    I have the same problem. I got the following working
    1. using http as in the media sample going to the internet
    2. local disk via file://
    3.local host via http:
    BTW: make sure you have all your codecs updated on XP (SP3) by using something like CCCP codec support. this caused me major headaches. as the media player kept on crashing if you pressed the start button more than once.
    I now need to show a stream via either RTP or UDP . my idea is too use the VLC player and stream the video and then use the media player to display the stream within my application
    I am using netbeans with javaFX1.2 XP with SP3.

  • Save a rtp stream to a file

    hi,
    I can send and receive a rtp stream. Now I want to save a received rtp stream to a file,but i got an exception when create
    a DataSink:
    public synchronized void update(ReceiveStreamEvent evt) {
            RTPManager mgr = (RTPManager) evt.getSource();
            Participant participant = evt.getParticipant();
            ReceiveStream stream = evt.getReceiveStream();
            if (evt instanceof NewReceiveStreamEvent) {
                DataSource ds = stream.getDataSource();
        }

    Well gee wiz, I should would like to help you with your exception with your data sink...
    But you didn't post what exceptino you were getting, so I can't help you...
    And the code you posted has nothing to do with a DataSink to begin with, so I can't help you again...

  • Writing a conference server for RTP streams

    Hello,
    I'm trying to write a conference server which accepts multiple RTP streams (one for each participant), creates a mixed RTP stream of all other participants and sends that stream back to each participant.
    For 2 participants, I was able to correctly receive and send the stream of the other participant to each party.
    For 3 participants, creating the merging data source does not seem to work - i.e. no data is received by the participants.
    I tried creating a cloneable data sources instead, thinking that this may be the root cause, but when creating cloneable data sources from incoming RTP sources, I am unable to get the Processor into Configured state, it seems to deadlock. Here's the code outline :
        Iterator pIt = participants.iterator();
        List dataSources = new ArrayList();
        while(pIt.hasNext()) {
          Party p = (Party) pIt.next();
          if(p!=dest) {
            DataSource ds = p.getDataSource();
            DataSource cds = Manager.createCloneableDataSource(ds);
            DataSource clone= ((SourceCloneable)cds).createClone();
            dataSources.add(clone);
        Object[] sources = dataSources.toArray(new DataSource[0]);
        DataSource dataSource =   Manager.createMergingDataSource((DataSource[])sources);
        Processor p = Manager.createProcessor(dataSource);
        MixControllerListener cl = new MixControllerListener();
        p.addControllerListener(cl);
        // Put the Processor into configured state.
        p.configure();
        if (!cl.waitForState(p, p.Configured)) {
            System.err.println("Failed to configure the processor.");
            assert false;
        }Here are couple of stack traces :
    "RTPEventHandler" daemon prio=1 tid=0x081d6828 nid=0x3ea6 in Object.wait() [98246000..98247238]
            at java.lang.Object.wait(Native Method)
            - waiting on <0x9f37e4a8> (a java.lang.Object)
            at java.lang.Object.wait(Object.java:429)
            at demo.Mixer$MixControllerListener.waitForState(Mixer.java:248)
            - locked <0x9f37e4a8> (a java.lang.Object)
            at demo.Mixer.createMergedDataSource(Mixer.java:202)
            at demo.Mixer.createSendStreams(Mixer.java:165)
            at demo.Mixer.createSendStreamsWhenAllJoined(Mixer.java:157)
            - locked <0x9f481840> (a demo.Mixer)
            at demo.Mixer.update(Mixer.java:123)
            at com.sun.media.rtp.RTPEventHandler.processEvent(RTPEventHandler.java:62)
            at com.sun.media.rtp.RTPEventHandler.dispatchEvents(RTPEventHandler.java:96)
            at com.sun.media.rtp.RTPEventHandler.run(RTPEventHandler.java:115)
    "JMF thread: com.sun.media.ProcessEngine@a3c5b6[ com.sun.media.ProcessEngine@a3c5b6 ] ( configureThread)" daemon prio=1 tid=0x082fe3c8 nid=0x3ea6 in Object.wait() [977e0000..977e1238]
            at java.lang.Object.wait(Native Method)
            - waiting on <0x9f387560> (a java.lang.Object)
            at java.lang.Object.wait(Object.java:429)
            at com.sun.media.parser.RawBufferParser$FrameTrack.parse(RawBufferParser.java:247)
            - locked <0x9f387560> (a java.lang.Object)
            at com.sun.media.parser.RawBufferParser.getTracks(RawBufferParser.java:112)
            at com.sun.media.BasicSourceModule.doRealize(BasicSourceModule.java:180)
            at com.sun.media.PlaybackEngine.doConfigure1(PlaybackEngine.java:229)
            at com.sun.media.ProcessEngine.doConfigure(ProcessEngine.java:43)
            at com.sun.media.ConfigureWorkThread.process(BasicController.java:1370)
            at com.sun.media.StateTransitionWorkThread.run(BasicController.java:1339)
    "JMF thread" daemon prio=1 tid=0x080db410 nid=0x3ea6 in Object.wait() [97f41000..97f41238]
            at java.lang.Object.wait(Native Method)
            - waiting on <0x9f480578> (a com.ibm.media.protocol.CloneableSourceStreamAdapter$PushBufferStreamSlave)
            at java.lang.Object.wait(Object.java:429)
            at com.ibm.media.protocol.CloneableSourceStreamAdapter$PushBufferStreamSlave.run(CloneableSourceStreamAdapter.java:375)
            - locked <0x9f480578> (a com.ibm.media.protocol.CloneableSourceStreamAdapter$PushBufferStreamSlave)
            at java.lang.Thread.run(Thread.java:534)Any ideas ?
    Thanks,
    Jarek

    bgl,
    I was able to get past the cloning issue by following the Clone.java example to the letter :)
    Turns out that the cloneable data source must be added as a send stream first, and then the clonet data source. Now for each party in the call the conf. server does the following :
    Party(RTPManager mgr,DataSource ds) {
          this.mgr=mgr;
          this.ds=Manager.createCloneableDataSource(ds);
       synchronized DataSource cloneDataSource() {
          DataSource retVal;
          if(getNeedsCloning()) {
            retVal = ((SourceCloneable) ds).createClone();
          } else {
            retVal = ds;
            setNeedsCloning();
          return retVal;
        private void setNeedsCloning() {
          needsCloning=true;
        private boolean getNeedsCloning() {
          return needsCloning;
         private synchronized void addSendStreamFromNewParticipant(Party newOne) throws UnsupportedFormatException, IOException {
        debug("*** - New one joined. Creating the send streams. Curr count :" + participants.size());
        Iterator pIt = participants.iterator();
        while(pIt.hasNext()) {
          Party p = (Party)pIt.next();
          assert p!=newOne;
          // update existing participant
          SendStream sendStream = p.getMgr().createSendStream(newOne.cloneDataSource(),0);
          sendStream.start();
          // send data from existing participant to the new one
          sendStream = newOne.getMgr().createSendStream(p.cloneDataSource(),0);
          sendStream.start();
        debug("*** - Done creating the streams.");So I made some progress, but I'm still not quite there.
    The RTP manager JavaDoc for createSendStream states the following :
    * This method is used to create a sending stream within the RTP
    * session. For each time the call is made, a new sending stream
    * will be created. This stream will use the SDES items as entered
    * in the initialize() call for all its RTCP messages. Each stream
    * is sent out with a new SSRC (Synchronisation SouRCe
    * identifier), but from the same participant i.e. local
    * participant. <BR>
    For 3 participants, my conf. server creates 2 send streams to every one of them, so I'd expect 2 SSRCs on the wire. Examining the RTP packets in Ethereal, I only see 1 SSRC, as if the 2nd createSendStream call failed. Consequently, each participany in the conference is able to receive voice from only 1 other participant, even though I create RTPManager instance for each participany, and add 2 send streams.
    Any ideas ?
    Thanks,
    Jarek

  • To retransmit the rtp streams received

    hi !
    Program A transmits media data to another program B using RTPSessionMgr.What B has to do is send the recived streams to another receiver program C.This has to play the stream.That is I am trying to implement a client router server model.
    So i maintained a session between A & B and also between B & C.Once NewReceiveStream event occurs at B,it retrieves the datasource from the stream and uses the createSendStream() method of the sessionmanager and sends to C.
    My problem is that,C receives the audio and video streams.but never plays it.I get a pink screen for video and no audio.
    can somebody help me out in pointing out my mistake or tell me a new strategy to implement this.
    Actually,i have tried to use only datagram sockets on the router side.ie I created 2 sockets for B to receive data. and forwarded to C using 2 other sockets .And this did work.But the client does not know the sender details.I require to maintain the sender,receiver reports.So i went for the session manager(RTPManager /RTPSessionMgr).
    kindly help.

    Hi all!
    Nice to meet ya, it very exciting that I got a project which will integrating the JMF and JXTA, it got RTP streams from the JMF framework and then sending it to the JXTA framework ,and sent to any peer in some groups to visualizing it .So some aspects of which is similar to yours ,would you mind add me to your contact list of MSN(my MSN account is:[email protected])

  • Receiving Video RTP Stream (JMF) in JME ( MMAPI ) - URGENT !!!

    Hi Folks...
    I�m trying to develop an application that sends the images from a web cam connected to the computer, to the PDA that the images can be viewed by the user...
    My code for the JMF RTP Video Stream is as follows.
    Processor proc = null;
              javax.media.protocol.DataSource ds = null;
              TrackControl[] tc = null;
              int y;
              boolean encodingOk = false;
              Vector<javax.media.protocol.DataSource> datasources = new Vector<javax.media.protocol.DataSource>();
              for( int x = 0 ; x < camerasInfo.length ; x++ ){
                   try {
                        proc = Manager.createProcessor(camerasInfo[x].getLocator());
                   }catch (NoProcessorException e) { System.out.println("Erro ao int�nciar PROCESSOR: " + e);     }
                    catch (IOException e) { System.out.println("Erro ao int�nciar PROCESSOR: " + e); }
                   proc.configure();
                   try {
                        Thread.sleep(2000);
                   } catch (InterruptedException e1) {
                        // TODO Auto-generated catch block
                        e1.printStackTrace();
                   proc.setContentDescriptor( new ContentDescriptor(ContentDescriptor.RAW_RTP) );
                   tc = proc.getTrackControls();
                   for( y = 0; y < tc.length ; y++ ){
                        if (!encodingOk && tc[y] instanceof FormatControl){
                             if( ( (FormatControl) tc[y] ).setFormat( new VideoFormat(VideoFormat.RGB) ) != null ){
                                  tc[y].setEnabled(true);                              
                             }else{
                                  encodingOk = true;
                        }else{
                             tc[y].setEnabled(false);
                   proc.realize();
                   try {
                        Thread.sleep(2000);
                   } catch (InterruptedException e1) {
                        e1.printStackTrace();
                   try{
                        ds = proc.getDataOutput();
                   }catch(NotRealizedError e){
                        System.out.println("ERRO ao realizar datasource: " + e);
                   }catch(ClassCastException e){
                        System.out.println("Erro ao realizar datasource: " + e);
                   datasources.add(ds);
                   System.out.println( ds.getLocator() );
                   encodingOk = false;
              MediaLocator ml = new MediaLocator("rtp://10.1.1.100:99/video");
              try {
                   DataSink datSink = Manager.createDataSink(ds , ml);
                   datSink.open();
                   datSink.start();
              } catch (NoDataSinkException e) {
                   System.out.println("Erro ao instanciar DataSink: " + e );
              } catch (SecurityException e) {
                   System.out.println("Erro ao iniciar DataSink: " + e);
              } catch (IOException e) {
                   System.out.println("Erro ao iniciar DataSink: " + e);
              }I�m not sure if this code is correctly... it is ?
    So... the next part of the systems runs on the PDA..
    The code that access this RTP Stream is as follows..
              VideoControl c = null;
              try {
                   player = Manager.createPlayer("rtp://10.1.1.100:99/video");
                   c = (VideoControl) player.getControl("VideoControl");
                   tela = (Item) c.initDisplayMode( GUIControl.USE_GUI_PRIMITIVE, null);
                   player.start();
                   append(tela);
              } catch (IOException e) {
                   str.setText(e.toString());
                   append( str );
              } catch (MediaException e) {
                   str.setText(e.toString());
                   append( str );
              }So when the APP try to create a player for "rtp://10.1.1.100:99/video" an MediaException is throwed..
    javax.microedition.media.MediaException: Unable to create Player for the locator: rtp://10.1.1.100:99/video
    So... I don�t know what happen =/
    The error is in the PDA module ? Or in the computer�s initialization off the RTP Video Streaming ?
    I need finish this job at next week... so any help is usefull..
    Waiting for answers
    Rodrigo Kerkhoff

    First of all: before going onto the j2me part, make sure the server works, before doing anything else! Apparently, it doesn't...
    The MediaLocator is generally used to specify a kind of URL depicting where the data put into the datasink should go to. In your case, This cannot be just some URL where you want it to act as a rtps server. You'll need to implement that server yourself I guess.

Maybe you are looking for

  • My Photos as Screen Saver will not 'save'

    When I select my own album or event to use as the screen saver for my new apple tv, it never seems to 'stick' or save. Everytime I come back into the room, or the next day it seems that the screensaver is reverting back to the built in 'animals' phot

  • Yahoo Search pops up each time I open a new page - Uninstall??

    Yahoo Search pops up each time I open a new page - Uninstall??

  • Forms 6i and Java

    Hi@all, is it possible to override the Forms Button, Textfield and so on with elements from the java awt class to make the forms mor appealing? Can you give me a short tutorial or somethig. Thank you for your help. greetings.

  • 11gR2 database portability from home and school

    I have a grad research project to refine a star datawarehouse started by another grad student working on this both at home and at the school. I will have a backup copy of the the test version which I can carry on a protable usb HD back and forth betw

  • Bad root Directory CRC

    Upon booting into Open Firmware, I get on the first line the message "Warning: bad Root Directory CRC" The rest of the lines that follow are normal. What is a bad Root Directory CRC and how does one fix it?