changeset 9565:e74916774667

Improved RTP packet buffering, by relying on the underlying OS's UDP socket buffering. Improve A/V sync by dropping packets when one stream gets too far behind the other. Now tries to figure out the video frame rate automatically (if "-fps" is not used). Added support for MPEG-4 Elementary Stream video and MPEG-4 Generic audio RTP streams.
author rsf
date Tue, 11 Mar 2003 19:08:31 +0000
parents 898e3692ca0d
children 015b404023f5
files libmpdemux/demux_rtp.cpp libmpdemux/demux_rtp_codec.cpp libmpdemux/demux_rtp_internal.h
diffstat 3 files changed, 348 insertions(+), 233 deletions(-) [+]
line wrap: on
line diff
--- a/libmpdemux/demux_rtp.cpp	Tue Mar 11 19:03:31 2003 +0000
+++ b/libmpdemux/demux_rtp.cpp	Tue Mar 11 19:08:31 2003 +0000
@@ -9,6 +9,7 @@
 
 #include "BasicUsageEnvironment.hh"
 #include "liveMedia.hh"
+#include "GroupsockHelper.hh"
 #include <unistd.h>
 
 extern "C" stream_t* stream_open_sdp(int fd, off_t fileSize,
@@ -43,41 +44,38 @@
   return 0;
 }
 
-// A data structure representing a buffer being read:
-class ReadBufferQueue; // forward
-class ReadBuffer {
-public:
-  ReadBuffer(ReadBufferQueue* ourQueue, demux_packet_t* dp);
-  virtual ~ReadBuffer();
-  Boolean enqueue();
-
-  demux_packet_t* dp() const { return fDP; }
-  ReadBufferQueue* ourQueue() { return fOurQueue; }
-
-  ReadBuffer* next;
-private:
-  demux_packet_t* fDP;
-  ReadBufferQueue* fOurQueue;
-};
-
+// A data structure representing input data for each stream:
 class ReadBufferQueue {
 public:
   ReadBufferQueue(MediaSubsession* subsession, demuxer_t* demuxer,
 		  char const* tag);
   virtual ~ReadBufferQueue();
 
-  ReadBuffer* dequeue();
-
   FramedSource* readSource() const { return fReadSource; }
   RTPSource* rtpSource() const { return fRTPSource; }
   demuxer_t* ourDemuxer() const { return fOurDemuxer; }
   char const* tag() const { return fTag; }
 
-  ReadBuffer* head;
-  ReadBuffer* tail;
   char blockingFlag; // used to implement synchronous reads
-  unsigned counter; // used for debugging
+
+  // For A/V synchronization:
+  Boolean prevPacketWasSynchronized;
+  float prevPacketPTS;
+  ReadBufferQueue** otherQueue;
+
+  // The 'queue' actually consists of just a single "demux_packet_t"
+  // (because the underlying OS does the actual queueing/buffering):
+  demux_packet_t* dp;
+
+  // However, we sometimes inspect buffers before delivering them.
+  // For this, we maintain a queue of pending buffers:
+  void savePendingBuffer(demux_packet_t* dp);
+  demux_packet_t* getPendingBuffer();
+
 private:
+  demux_packet_t* pendingDPHead;
+  demux_packet_t* pendingDPTail;
+
   FramedSource* fReadSource;
   RTPSource* fRTPSource;
   demuxer_t* fOurDemuxer;
@@ -99,10 +97,6 @@
 int rtspStreamOverTCP = 0; 
 
 extern "C" void demux_open_rtp(demuxer_t* demuxer) {
-  if (rtspStreamOverTCP && LIVEMEDIA_LIBRARY_VERSION_INT < 1033689600) {
-    fprintf(stderr, "TCP streaming of RTP/RTCP requires \"LIVE.COM Streaming Media\" library version 2002.10.04 or later - ignoring the \"-rtsp-stream-over-tcp\" flag\n");
-    rtspStreamOverTCP = 0;
-  }
   do {
     TaskScheduler* scheduler = BasicTaskScheduler::createNew();
     if (scheduler == NULL) break;
@@ -110,7 +104,6 @@
     if (env == NULL) break;
 
     RTSPClient* rtspClient = NULL;
-    unsigned flags = 0;
 
     if (demuxer == NULL || demuxer->stream == NULL) break;  // shouldn't happen
     demuxer->stream->eof = 0; // just in case 
@@ -120,7 +113,7 @@
     char* sdpDescription = (char*)(demuxer->stream->priv);
     if (sdpDescription == NULL) {
       // We weren't given a SDP description directly, so assume that
-      // we were give a RTSP URL
+      // we were given a RTSP URL:
       char const* url = demuxer->stream->streaming_ctrl->url->url;
 
       extern int verbose;
@@ -151,19 +144,20 @@
     rtpState->rtspClient = rtspClient;
     rtpState->mediaSession = mediaSession;
     rtpState->audioBufferQueue = rtpState->videoBufferQueue = NULL;
+    rtpState->flags = 0;
     rtpState->firstSyncTime.tv_sec = rtpState->firstSyncTime.tv_usec = 0;
     demuxer->priv = rtpState;
 
     // Create RTP receivers (sources) for each subsession:
     MediaSubsessionIterator iter(*mediaSession);
     MediaSubsession* subsession;
-    unsigned streamType = 0; // 0 => video; 1 => audio
+    unsigned desiredReceiveBufferSize;
     while ((subsession = iter.next()) != NULL) {
       // Ignore any subsession that's not audio or video:
       if (strcmp(subsession->mediumName(), "audio") == 0) {
-	streamType = 1;
+	desiredReceiveBufferSize = 100000;
       } else if (strcmp(subsession->mediumName(), "video") == 0) {
-	streamType = 0;
+	desiredReceiveBufferSize = 2000000;
       } else {
 	continue;
       }
@@ -173,27 +167,52 @@
       } else {
 	fprintf(stderr, "Initiated \"%s/%s\" RTP subsession\n", subsession->mediumName(), subsession->codecName());
 
-	if (rtspClient != NULL) {
-	  // Issue RTSP "SETUP" and "PLAY" commands on the chosen subsession:
-	  if (!rtspClient->setupMediaSubsession(*subsession, False,
-						rtspStreamOverTCP)) break;
-	  if (!rtspClient->playMediaSubsession(*subsession)) break;
+	// Set the OS's socket receive buffer sufficiently large to avoid
+	// incoming packets getting dropped between successive reads from this
+	// subsession's demuxer.  Depending on the bitrate(s) that you expect,
+	// you may wish to tweak the "desiredReceiveBufferSize" values above.
+	int rtpSocketNum = subsession->rtpSource()->RTPgs()->socketNum();
+	int receiveBufferSize
+	  = increaseReceiveBufferTo(*env, rtpSocketNum,
+				    desiredReceiveBufferSize);
+	if (verbose > 0) {
+	  fprintf(stderr, "Increased %s socket receive buffer to %d bytes \n",
+		  subsession->mediumName(), receiveBufferSize);
 	}
 
-	// Now that the subsession is ready to be read, do additional
-	// MPlayer codec-specific initialization on it:
-	if (streamType == 0) { // video
-	  rtpState->videoBufferQueue
-	    = new ReadBufferQueue(subsession, demuxer, "video");
-	  rtpCodecInitialize_video(demuxer, subsession, flags);
-	} else { // audio
-	  rtpState->audioBufferQueue
-	    = new ReadBufferQueue(subsession, demuxer, "audio");
-	  rtpCodecInitialize_audio(demuxer, subsession, flags);
+	if (rtspClient != NULL) {
+	  // Issue a RTSP "SETUP" command on the chosen subsession:
+	  if (!rtspClient->setupMediaSubsession(*subsession, False,
+						rtspStreamOverTCP)) break;
 	}
       }
     }
-    rtpState->flags = flags;
+
+    if (rtspClient != NULL) {
+      // Issue a RTSP aggregate "PLAY" command on the whole session:
+      if (!rtspClient->playMediaSession(*mediaSession)) break;
+    }
+
+    // Now that the session is ready to be read, do additional
+    // MPlayer codec-specific initialization on each subsession:
+    iter.reset();
+    while ((subsession = iter.next()) != NULL) {
+      if (subsession->readSource() == NULL) continue; // not reading this
+
+      unsigned flags = 0;
+      if (strcmp(subsession->mediumName(), "audio") == 0) {
+	rtpState->audioBufferQueue
+	  = new ReadBufferQueue(subsession, demuxer, "audio");
+	rtpState->audioBufferQueue->otherQueue = &(rtpState->videoBufferQueue);
+	rtpCodecInitialize_audio(demuxer, subsession, flags);
+      } else if (strcmp(subsession->mediumName(), "video") == 0) {
+	rtpState->videoBufferQueue
+	  = new ReadBufferQueue(subsession, demuxer, "video");
+	rtpState->videoBufferQueue->otherQueue = &(rtpState->audioBufferQueue);
+	rtpCodecInitialize_video(demuxer, subsession, flags);
+      }
+      rtpState->flags |= flags;
+    }
   } while (0);
 }
 
@@ -201,11 +220,12 @@
   // Get the RTP state that was stored in the demuxer's 'priv' field:
   RTPState* rtpState = (RTPState*)(demuxer->priv);
 
-  return (rtpState->flags&RTPSTATE_IS_MPEG) != 0;
+  return (rtpState->flags&RTPSTATE_IS_MPEG12_VIDEO) != 0;
 }
 
-static ReadBuffer* getBuffer(ReadBufferQueue* bufferQueue,
-			     demuxer_t* demuxer); // forward
+static demux_packet_t* getBuffer(demuxer_t* demuxer, demux_stream_t* ds,
+				 Boolean mustGetNewData,
+				 float& ptsBehind); // forward
 
 extern "C" int demux_rtp_fill_buffer(demuxer_t* demuxer, demux_stream_t* ds) {
   // Get a filled-in "demux_packet" from the RTP source, and deliver it.
@@ -213,7 +233,51 @@
   // to block in the (hopefully infrequent) case where no packet is
   // immediately available.
 
-  // Begin by finding the buffer queue that we want to read from:
+  while (1) {
+    float ptsBehind;
+    demux_packet_t* dp = getBuffer(demuxer, ds, False, ptsBehind); // blocking
+    if (dp == NULL) return 0;
+
+    if (demuxer->stream->eof) return 0; // source stream has closed down
+  
+    // Before using this packet, check to make sure that its presentation
+    // time is not far behind the other stream (if any).  If it is,
+    // then we discard this packet, and get another instead.  (The rest of
+    // MPlayer doesn't always do a good job of synchronizing when the
+    // audio and video streams get this far apart.)
+    // (We don't do this when streaming over TCP, because then the audio and
+    // video streams are interleaved.)
+    const float ptsBehindThreshold = 1.0; // seconds
+    if (ptsBehind < ptsBehindThreshold || rtspStreamOverTCP) { // packet's OK
+      ds_add_packet(ds, dp);
+      break;
+    }
+    
+    free_demux_packet(dp); // give back this packet, and get another one
+  }
+
+  return 1;
+}
+
+Boolean awaitRTPPacket(demuxer_t* demuxer, demux_stream_t* ds,
+		       unsigned char*& packetData, unsigned& packetDataLen,
+		       float& pts) {
+  // Similar to "demux_rtp_fill_buffer()", except that the "demux_packet"
+  // is not delivered to the "demux_stream".
+  float ptsBehind;
+  demux_packet_t* dp = getBuffer(demuxer, ds, True, ptsBehind); // blocking
+  if (dp == NULL) return False;
+
+  packetData = dp->buffer;
+  packetDataLen = dp->len;
+  pts = dp->pts;
+
+  return True;
+}
+
+Boolean insertRTPData(demuxer_t* demuxer, demux_stream_t* ds,
+		      unsigned char* data, unsigned dataLen) {
+  // Begin by finding the buffer queue that we want to add data to.
   // (Get this from the RTP state, which we stored in
   //  the demuxer's 'priv' field)
   RTPState* rtpState = (RTPState*)(demuxer->priv);
@@ -223,54 +287,23 @@
   } else if (ds == demuxer->audio) {
     bufferQueue = rtpState->audioBufferQueue;
   } else {
-    fprintf(stderr, "demux_rtp_fill_buffer: internal error: unknown stream\n");
-    return 0;
-  }
-
-  if (bufferQueue == NULL || bufferQueue->readSource() == NULL) {
-    fprintf(stderr, "demux_rtp_fill_buffer failed: no appropriate RTP subsession has been set up\n");
-    return 0;
-  }
-  
-  ReadBuffer* readBuffer = getBuffer(bufferQueue, demuxer); // blocking
-  if (readBuffer != NULL) ds_add_packet(ds, readBuffer->dp());
-
-  if (demuxer->stream->eof) return 0; // source stream has closed down
-
-  return 1;
-}
-
-Boolean awaitRTPPacket(demuxer_t* demuxer, unsigned streamType,
-		       unsigned char*& packetData, unsigned& packetDataLen) {
-  // Begin by finding the buffer queue that we want to read from:
-  // (Get this from the RTP state, which we stored in
-  //  the demuxer's 'priv' field)
-  RTPState* rtpState = (RTPState*)(demuxer->priv);
-  ReadBufferQueue* bufferQueue = NULL;
-  if (streamType == 0) {
-    bufferQueue = rtpState->videoBufferQueue;
-  } else if (streamType == 1) {
-    bufferQueue = rtpState->audioBufferQueue;
-  } else {
-    fprintf(stderr, "awaitRTPPacket: internal error: unknown streamType %d\n",
-	    streamType);
+    fprintf(stderr, "(demux_rtp)insertRTPData: internal error: unknown stream\n");
     return False;
   }
 
-  if (bufferQueue == NULL || bufferQueue->readSource() == NULL) {
-    fprintf(stderr, "awaitRTPPacket failed: no appropriate RTP subsession has been set up\n");
-    return False;
-  }
-  
-  ReadBuffer* readBuffer = getBuffer(bufferQueue, demuxer); // blocking
-  if (readBuffer == NULL) return False;
+  if (data == NULL || dataLen == 0) return False;
+
+  demux_packet_t* dp = new_demux_packet(dataLen);
+  if (dp == NULL) return False;
 
-  demux_packet_t* dp = readBuffer->dp();
-  packetData = dp->buffer;
-  packetDataLen = dp->len;
+  // Copy our data into the buffer, and save it:
+  memmove(dp->buffer, data, dataLen);
+  dp->len = dataLen;
+  dp->pts = 0;
+  bufferQueue->savePendingBuffer(dp);
+}
 
-  return True;
-}
+static void teardownRTSPSession(RTPState* rtpState); // forward
 
 extern "C" void demux_close_rtp(demuxer_t* demuxer) {
   // Reclaim all RTP-related state:
@@ -278,6 +311,9 @@
   // Get the RTP state that was stored in the demuxer's 'priv' field:
   RTPState* rtpState = (RTPState*)(demuxer->priv);
   if (rtpState == NULL) return;
+
+  teardownRTSPSession(rtpState);
+
   UsageEnvironment* env = NULL;
   TaskScheduler* scheduler = NULL;
   if (rtpState->mediaSession != NULL) {
@@ -296,76 +332,65 @@
 
 ////////// Extra routines that help implement the above interface functions:
 
-static void afterReading(void* clientData, unsigned frameSize,
-			 struct timeval presentationTime); // forward
-static void onSourceClosure(void* clientData); // forward
-
-static void scheduleNewBufferRead(ReadBufferQueue* bufferQueue) {
-  if (bufferQueue->readSource()->isCurrentlyAwaitingData()) return;
-      // a read from this source is already in progress
-
-  // Allocate a new packet buffer, and arrange to read into it:
-  unsigned const bufferSize = 30000; // >= the largest conceivable RTP packet
-  demux_packet_t* dp = new_demux_packet(bufferSize);
-  if (dp == NULL) return;
-  ReadBuffer* readBuffer = new ReadBuffer(bufferQueue, dp);
-
-  // Schedule the read operation:
-  bufferQueue->readSource()->getNextFrame(dp->buffer, bufferSize,
-					  afterReading, readBuffer,
-					  onSourceClosure, readBuffer);
-}
+#define MAX_RTP_FRAME_SIZE 50000
+    // >= the largest conceivable frame composed from one or more RTP packets
 
 static void afterReading(void* clientData, unsigned frameSize,
 			 struct timeval presentationTime) {
-  ReadBuffer* readBuffer = (ReadBuffer*)clientData;
-  ReadBufferQueue* bufferQueue = readBuffer->ourQueue();
+  if (frameSize >= MAX_RTP_FRAME_SIZE) {
+    fprintf(stderr, "Saw an input frame too large (>=%d).  Increase MAX_RTP_FRAME_SIZE in \"demux_rtp.cpp\".\n",
+	    MAX_RTP_FRAME_SIZE);
+  }
+  ReadBufferQueue* bufferQueue = (ReadBufferQueue*)clientData;
   demuxer_t* demuxer = bufferQueue->ourDemuxer();
   RTPState* rtpState = (RTPState*)(demuxer->priv);
 
   if (frameSize > 0) demuxer->stream->eof = 0;
 
-  demux_packet_t* dp = readBuffer->dp();
+  demux_packet_t* dp = bufferQueue->dp;
   dp->len = frameSize;
 
   // Set the packet's presentation time stamp, depending on whether or
   // not our RTP source's timestamps have been synchronized yet: 
-  {
-    Boolean hasBeenSynchronized
-      = bufferQueue->rtpSource()->hasBeenSynchronizedUsingRTCP();
-    if (hasBeenSynchronized) {
-      struct timeval* fst = &(rtpState->firstSyncTime); // abbrev
-      if (fst->tv_sec == 0 && fst->tv_usec == 0) {
-	*fst = presentationTime;
-      }
+  Boolean hasBeenSynchronized
+    = bufferQueue->rtpSource()->hasBeenSynchronizedUsingRTCP();
+  if (hasBeenSynchronized) {
+    if (verbose > 0 && !bufferQueue->prevPacketWasSynchronized) {
+      fprintf(stderr, "%s stream has been synchronized using RTCP \n",
+	      bufferQueue->tag());
+    }
 
-      // For the "pts" field, use the time differential from the first
-      // synchronized time, rather than absolute time, in order to avoid
-      // round-off errors when converting to a float:
-      dp->pts = presentationTime.tv_sec - fst->tv_sec
-	+ (presentationTime.tv_usec - fst->tv_usec)/1000000.0;
-    } else {
-      dp->pts = 0.0;
+    struct timeval* fst = &(rtpState->firstSyncTime); // abbrev
+    if (fst->tv_sec == 0 && fst->tv_usec == 0) {
+      *fst = presentationTime;
     }
+    
+    // For the "pts" field, use the time differential from the first
+    // synchronized time, rather than absolute time, in order to avoid
+    // round-off errors when converting to a float:
+    dp->pts = presentationTime.tv_sec - fst->tv_sec
+      + (presentationTime.tv_usec - fst->tv_usec)/1000000.0;
+    bufferQueue->prevPacketPTS = dp->pts;
+  } else {
+    if (verbose > 0 && bufferQueue->prevPacketWasSynchronized) {
+      fprintf(stderr, "%s stream is no longer RTCP-synchronized \n",
+	      bufferQueue->tag());
+    }
+
+    // use the previous packet's "pts" once again:
+    dp->pts = bufferQueue->prevPacketPTS;
   }
+  bufferQueue->prevPacketWasSynchronized = hasBeenSynchronized;
 
   dp->pos = demuxer->filepos;
   demuxer->filepos += frameSize;
-  if (!readBuffer->enqueue()) {
-    // The queue is full, so discard the buffer:
-    delete readBuffer;
-  }
 
   // Signal any pending 'doEventLoop()' call on this queue:
   bufferQueue->blockingFlag = ~0;
-
-  // Finally, arrange to do another read, if appropriate
-  scheduleNewBufferRead(bufferQueue);
 }
 
 static void onSourceClosure(void* clientData) {
-  ReadBuffer* readBuffer = (ReadBuffer*)clientData;
-  ReadBufferQueue* bufferQueue = readBuffer->ourQueue();
+  ReadBufferQueue* bufferQueue = (ReadBufferQueue*)clientData;
   demuxer_t* demuxer = bufferQueue->ourDemuxer();
 
   demuxer->stream->eof = 1;
@@ -374,90 +399,123 @@
   bufferQueue->blockingFlag = ~0;
 }
 
-static ReadBuffer* getBufferIfAvailable(ReadBufferQueue* bufferQueue) {
-  ReadBuffer* readBuffer = bufferQueue->dequeue();
+static demux_packet_t* getBuffer(demuxer_t* demuxer, demux_stream_t* ds,
+				 Boolean mustGetNewData,
+				 float& ptsBehind) {
+  // Begin by finding the buffer queue that we want to read from:
+  // (Get this from the RTP state, which we stored in
+  //  the demuxer's 'priv' field)
+  RTPState* rtpState = (RTPState*)(demuxer->priv);
+  ReadBufferQueue* bufferQueue = NULL;
+  if (ds == demuxer->video) {
+    bufferQueue = rtpState->videoBufferQueue;
+  } else if (ds == demuxer->audio) {
+    bufferQueue = rtpState->audioBufferQueue;
+  } else {
+    fprintf(stderr, "(demux_rtp)getBuffer: internal error: unknown stream\n");
+    return NULL;
+  }
+
+  if (bufferQueue == NULL || bufferQueue->readSource() == NULL) {
+    fprintf(stderr, "(demux_rtp)getBuffer failed: no appropriate RTP subsession has been set up\n");
+    return NULL;
+  }
+  
+  demux_packet_t* dp;
+  if (!mustGetNewData) {
+    // Check whether we have a previously-saved buffer that we can use:
+    dp = bufferQueue->getPendingBuffer();
+    if (dp != NULL) return dp;
+  }
 
-  // Arrange to read a new packet into this queue:
-  scheduleNewBufferRead(bufferQueue);
+  // Allocate a new packet buffer, and arrange to read into it:
+  dp = new_demux_packet(MAX_RTP_FRAME_SIZE);
+  bufferQueue->dp = dp;
+  if (dp == NULL) return NULL;
+
+  // Schedule the read operation:
+  bufferQueue->blockingFlag = 0;
+  bufferQueue->readSource()->getNextFrame(dp->buffer, MAX_RTP_FRAME_SIZE,
+					  afterReading, bufferQueue,
+					  onSourceClosure, bufferQueue);
+  // Block ourselves until data becomes available:
+  TaskScheduler& scheduler
+    = bufferQueue->readSource()->envir().taskScheduler();
+  scheduler.doEventLoop(&bufferQueue->blockingFlag);
 
-  return readBuffer;
+  // Set the "ptsBehind" result parameter:
+  if (bufferQueue->prevPacketPTS != 0.0 && *(bufferQueue->otherQueue) != NULL
+      && (*(bufferQueue->otherQueue))->prevPacketPTS != 0.0) {
+    ptsBehind = (*(bufferQueue->otherQueue))->prevPacketPTS
+		 - bufferQueue->prevPacketPTS;
+  } else {
+    ptsBehind = 0.0;
+  }
+
+  if (mustGetNewData) {
+    // Save this buffer for future reads:
+    bufferQueue->savePendingBuffer(dp);
+  }
+
+  return dp;
 }
 
-static ReadBuffer* getBuffer(ReadBufferQueue* bufferQueue,
-			     demuxer_t* demuxer) {
-  // Check whether there's a full buffer to deliver to the client:
-  bufferQueue->blockingFlag = 0;
-  ReadBuffer* readBuffer;
-  while ((readBuffer = getBufferIfAvailable(bufferQueue)) == NULL
-	 && !demuxer->stream->eof) {
-    // Because we weren't able to deliver a buffer to the client immediately,
-    // block myself until one comes available:
-    TaskScheduler& scheduler
-      = bufferQueue->readSource()->envir().taskScheduler();
-#if USAGEENVIRONMENT_LIBRARY_VERSION_INT >= 1038614400
-    scheduler.doEventLoop(&bufferQueue->blockingFlag);
-#else
-    scheduler.blockMyself(&bufferQueue->blockingFlag);
-#endif
+static void teardownRTSPSession(RTPState* rtpState) {
+  RTSPClient* rtspClient = rtpState->rtspClient;
+  MediaSession* mediaSession = rtpState->mediaSession;
+  if (rtspClient == NULL || mediaSession == NULL) return;
+
+  MediaSubsessionIterator iter(*mediaSession);
+  MediaSubsession* subsession;
+
+  while ((subsession = iter.next()) != NULL) {
+    rtspClient->teardownMediaSubsession(*subsession);
   }
-
-  return readBuffer;
 }
 
 ////////// "ReadBuffer" and "ReadBufferQueue" implementation:
 
-#define MAX_QUEUE_SIZE 5
-
-ReadBuffer::ReadBuffer(ReadBufferQueue* ourQueue, demux_packet_t* dp)
-  : next(NULL), fDP(dp), fOurQueue(ourQueue) {
-}
-
-Boolean ReadBuffer::enqueue() {
-  if (fOurQueue->counter >= MAX_QUEUE_SIZE) {
-    // This queue is full.  Clear out an old entry from it, so that
-    // this new one will fit:
-    while (fOurQueue->counter >= MAX_QUEUE_SIZE) {
-      delete fOurQueue->dequeue();
-    }
-  }
-
-  // Add ourselves to the tail of our queue:
-  if (fOurQueue->tail == NULL) {
-    fOurQueue->head = this;
-  } else {
-    fOurQueue->tail->next = this;
-  }
-  fOurQueue->tail = this;
-  ++fOurQueue->counter;
-
-  return True;
-}
-
-ReadBuffer::~ReadBuffer() {
-  free_demux_packet(fDP);
-  delete next;
-}
-
 ReadBufferQueue::ReadBufferQueue(MediaSubsession* subsession,
 				 demuxer_t* demuxer, char const* tag)
-  : head(NULL), tail(NULL), counter(0),
+  : prevPacketWasSynchronized(False), prevPacketPTS(0.0), otherQueue(NULL),
+    dp(NULL), pendingDPHead(NULL), pendingDPTail(NULL),
     fReadSource(subsession == NULL ? NULL : subsession->readSource()),
     fRTPSource(subsession == NULL ? NULL : subsession->rtpSource()),
     fOurDemuxer(demuxer), fTag(strdup(tag)) {
 } 
 
 ReadBufferQueue::~ReadBufferQueue() {
-  delete head;
   delete fTag;
+
+  // Free any pending buffers (that never got delivered):
+  demux_packet_t* dp = pendingDPHead;
+  while (dp != NULL) {
+    demux_packet_t* dpNext = dp->next;
+    dp->next = NULL;
+    free_demux_packet(dp);
+    dp = dpNext;
+  }
 }
 
-ReadBuffer* ReadBufferQueue::dequeue() {
-  ReadBuffer* readBuffer = head;
-  if (readBuffer != NULL) {
-    head = readBuffer->next;
-    if (head == NULL) tail = NULL; 
-    --counter;
-    readBuffer->next = NULL;
+void ReadBufferQueue::savePendingBuffer(demux_packet_t* dp) {
+  // Keep this buffer around, until MPlayer asks for it later:
+  if (pendingDPTail == NULL) {
+    pendingDPHead = pendingDPTail = dp;
+  } else {
+    pendingDPTail->next = dp;
+    pendingDPTail = dp;
   }
-  return readBuffer;
+  dp->next = NULL;
 }
+
+demux_packet_t* ReadBufferQueue::getPendingBuffer() {
+  demux_packet_t* dp = pendingDPHead;
+  if (dp != NULL) {
+    pendingDPHead = dp->next;
+    if (pendingDPHead == NULL) pendingDPTail = NULL; 
+
+    dp->next = NULL;
+  }
+
+  return dp;
+}
--- a/libmpdemux/demux_rtp_codec.cpp	Tue Mar 11 19:03:31 2003 +0000
+++ b/libmpdemux/demux_rtp_codec.cpp	Tue Mar 11 19:08:31 2003 +0000
@@ -6,6 +6,8 @@
 #include "stheader.h"
 }
 
+static void
+needVideoFrameRate(demuxer_t* demuxer, MediaSubsession* subsession); // forward
 static Boolean
 parseQTState_video(QuickTimeGenericRTPSource::QTState const& qtState,
 		   unsigned& fourcc); // forward
@@ -27,35 +29,38 @@
   demux_stream_t* d_video = demuxer->video;
   d_video->sh = sh_video; sh_video->ds = d_video;
   
-  // If we happen to know the subsession's video frame rate, set it,
-  // so that the user doesn't have to give the "-fps" option instead.
-  int fps = (int)(subsession->videoFPS());
-  if (fps != 0) sh_video->fps = fps;
-  
   // Map known video MIME types to the BITMAPINFOHEADER parameters
   // that this program uses.  (Note that not all types need all
   // of the parameters to be set.)
   if (strcmp(subsession->codecName(), "MPV") == 0 ||
       strcmp(subsession->codecName(), "MP1S") == 0 ||
       strcmp(subsession->codecName(), "MP2T") == 0) {
-    flags |= RTPSTATE_IS_MPEG;
+    flags |= RTPSTATE_IS_MPEG12_VIDEO;
   } else if (strcmp(subsession->codecName(), "H263") == 0 ||
 	     strcmp(subsession->codecName(), "H263-1998") == 0) {
     bih->biCompression = sh_video->format
       = mmioFOURCC('H','2','6','3');
+    needVideoFrameRate(demuxer, subsession);
   } else if (strcmp(subsession->codecName(), "H261") == 0) {
     bih->biCompression = sh_video->format
       = mmioFOURCC('H','2','6','1');
+    needVideoFrameRate(demuxer, subsession);
   } else if (strcmp(subsession->codecName(), "JPEG") == 0) {
     bih->biCompression = sh_video->format
       = mmioFOURCC('M','J','P','G');
-#if (LIVEMEDIA_LIBRARY_VERSION_INT < 1044662400)
-    fprintf(stderr, "WARNING: This video stream might not play correctly.  Please upgrade to version \"2003.02.08\" or later of the \"LIVE.COM Streaming Media\" libraries.\n");
-#endif
+    needVideoFrameRate(demuxer, subsession);
   } else if (strcmp(subsession->codecName(), "MP4V-ES") == 0) {
     bih->biCompression = sh_video->format
       = mmioFOURCC('m','p','4','v');
-    //flags |= RTPSTATE_IS_MPEG; // MPEG hdr checking in video.c doesn't work!
+    // For the codec to work correctly, it may need a 'VOL Header' to be
+    // inserted at the front of the data stream.  Construct this from the
+    // "config" MIME parameter, which was present (hopefully) in the
+    // session's SDP description:
+    unsigned configLen;
+    unsigned char* configData
+      = parseGeneralConfigStr(subsession->fmtp_config(), configLen);
+    insertRTPData(demuxer, demuxer->video, configData, configLen);
+    needVideoFrameRate(demuxer, subsession);
   } else if (strcmp(subsession->codecName(), "X-QT") == 0 ||
 	     strcmp(subsession->codecName(), "X-QUICKTIME") == 0) {
     // QuickTime generic RTP format, as described in
@@ -64,12 +69,13 @@
     // We can't initialize this stream until we've received the first packet
     // that has QuickTime "sdAtom" information in the header.  So, keep
     // reading packets until we get one:
-    unsigned char* packetData; unsigned packetDataLen;
+    unsigned char* packetData; unsigned packetDataLen; float pts;
     QuickTimeGenericRTPSource* qtRTPSource
       = (QuickTimeGenericRTPSource*)(subsession->rtpSource());
     unsigned fourcc;
     do {
-      if (!awaitRTPPacket(demuxer, 0 /*video*/, packetData, packetDataLen)) {
+      if (!awaitRTPPacket(demuxer, demuxer->video,
+			  packetData, packetDataLen, pts)) {
 	return;
       }
     } while (!parseQTState_video(qtRTPSource->qtState, fourcc));
@@ -94,6 +100,8 @@
   demux_stream_t* d_audio = demuxer->audio;
   d_audio->sh = sh_audio; sh_audio->ds = d_audio;
   
+  wf->nChannels = subsession->numChannels();
+
   // Map known audio MIME types to the WAVEFORMATEX parameters
   // that this program uses.  (Note that not all types need all
   // of the parameters to be set.)
@@ -105,44 +113,35 @@
     wf->wFormatTag = sh_audio->format = 0x55;
     // Note: 0x55 is for layer III, but should work for I,II also
     wf->nSamplesPerSec = 0; // sample rate is deduced from the data
-    flags |= RTPSTATE_IS_MPEG;
   } else if (strcmp(subsession->codecName(), "AC3") == 0) {
     wf->wFormatTag = sh_audio->format = 0x2000;
     wf->nSamplesPerSec = 0; // sample rate is deduced from the data
   } else if (strcmp(subsession->codecName(), "PCMU") == 0) {
     wf->wFormatTag = sh_audio->format = 0x7;
-    wf->nChannels = 1;
     wf->nAvgBytesPerSec = 8000;
     wf->nBlockAlign = 1;
     wf->wBitsPerSample = 8;
     wf->cbSize = 0;
   } else if (strcmp(subsession->codecName(), "PCMA") == 0) {
     wf->wFormatTag = sh_audio->format = 0x6;
-    wf->nChannels = 1;
     wf->nAvgBytesPerSec = 8000;
     wf->nBlockAlign = 1;
     wf->wBitsPerSample = 8;
     wf->cbSize = 0;
   } else if (strcmp(subsession->codecName(), "GSM") == 0) {
     wf->wFormatTag = sh_audio->format = mmioFOURCC('a','g','s','m');
-    wf->nChannels = 1;
     wf->nAvgBytesPerSec = 1650;
     wf->nBlockAlign = 33;
     wf->wBitsPerSample = 16;
     wf->cbSize = 0;
   } else if (strcmp(subsession->codecName(), "QCELP") == 0) {
     wf->wFormatTag = sh_audio->format = mmioFOURCC('Q','c','l','p');
-    // The following settings for QCELP don't quite work right #####
-    wf->nChannels = 1;
     wf->nAvgBytesPerSec = 1750;
     wf->nBlockAlign = 35;
     wf->wBitsPerSample = 16;
     wf->cbSize = 0;
   } else if (strcmp(subsession->codecName(), "MP4A-LATM") == 0) {
     wf->wFormatTag = sh_audio->format = mmioFOURCC('m','p','4','a');
-#if (LIVEMEDIA_LIBRARY_VERSION_INT < 1042761600)
-    fprintf(stderr, "WARNING: This audio stream might not play correctly.  Please upgrade to version \"2003.01.17\" or later of the \"LIVE.COM Streaming Media\" libraries.\n");
-#else
     // For the codec to work correctly, it needs "AudioSpecificConfig"
     // data, which is parsed from the "StreamMuxConfig" string that
     // was present (hopefully) in the SDP description:
@@ -151,8 +150,15 @@
       = parseStreamMuxConfigStr(subsession->fmtp_config(),
 				codecdata_len);
     sh_audio->codecdata_len = codecdata_len;
-#endif
-    flags |= RTPSTATE_IS_MPEG;
+  } else if (strcmp(subsession->codecName(), "MPEG4-GENERIC") == 0) {
+    wf->wFormatTag = sh_audio->format = mmioFOURCC('m','p','4','a');
+    // For the codec to work correctly, it needs "AudioSpecificConfig"
+    // data, which was present (hopefully) in the SDP description:
+    unsigned codecdata_len;
+    sh_audio->codecdata
+      = parseGeneralConfigStr(subsession->fmtp_config(),
+			      codecdata_len);
+    sh_audio->codecdata_len = codecdata_len;
   } else if (strcmp(subsession->codecName(), "X-QT") == 0 ||
 	     strcmp(subsession->codecName(), "X-QUICKTIME") == 0) {
     // QuickTime generic RTP format, as described in
@@ -161,12 +167,13 @@
     // We can't initialize this stream until we've received the first packet
     // that has QuickTime "sdAtom" information in the header.  So, keep
     // reading packets until we get one:
-    unsigned char* packetData; unsigned packetDataLen;
+    unsigned char* packetData; unsigned packetDataLen; float pts;
     QuickTimeGenericRTPSource* qtRTPSource
       = (QuickTimeGenericRTPSource*)(subsession->rtpSource());
     unsigned fourcc, numChannels;
     do {
-      if (!awaitRTPPacket(demuxer, 1 /*audio*/, packetData, packetDataLen)) {
+      if (!awaitRTPPacket(demuxer, demuxer->audio,
+			  packetData, packetDataLen, pts)) {
 	return;
       }
     } while (!parseQTState_audio(qtRTPSource->qtState, fourcc, numChannels));
@@ -180,6 +187,47 @@
   }
 }
 
+static void needVideoFrameRate(demuxer_t* demuxer,
+			       MediaSubsession* subsession) {
+  // For some codecs, MPlayer's decoding software can't (or refuses to :-)
+  // figure out the frame rate by itself, so (unless the user specifies
+  // it manually, using "-fps") we figure it out ourselves here, using the
+  // presentation timestamps in successive packets,
+  extern float force_fps; if (force_fps != 0.0) return; // user used "-fps"
+
+  demux_stream_t* d_video = demuxer->video;
+  sh_video_t* sh_video = (sh_video_t*)(demuxer->video->sh);
+
+  // If we already know the subsession's video frame rate, use it:
+  int fps = (int)(subsession->videoFPS());
+  if (fps != 0) {
+    sh_video->fps = fps;
+    return;
+  }
+  
+  // Keep looking at incoming frames until we see two with different,
+  // non-zero "pts" timestamps:
+  unsigned char* packetData; unsigned packetDataLen;
+  float lastPTS = 0.0, curPTS;
+  unsigned const maxNumFramesToWaitFor = 100;
+  for (unsigned i = 0; i < maxNumFramesToWaitFor; ++i) {
+    if (!awaitRTPPacket(demuxer, demuxer->video,
+			packetData, packetDataLen, curPTS)) break;
+
+    if (curPTS > lastPTS && lastPTS != 0.0) {
+      // Use the difference between these two "pts"s to guess the frame rate.
+      // (should really check that there were no missing frames inbetween)#####
+      // Guess the frame rate as an integer.  If it's not, use "-fps" instead.
+      fps = (int)(1/(curPTS-lastPTS) + 0.5); // rounding
+      fprintf(stderr, "demux_rtp: Guessed the video frame rate as %d frames-per-second.\n\t(If this is wrong, use the \"-fps <frame-rate>\" option instead.)\n", fps);
+      sh_video->fps = fps;
+      return;
+    }
+    lastPTS = curPTS;
+  }
+  fprintf(stderr, "demux_rtp: Failed to guess the video frame rate\n");
+}
+
 static Boolean
 parseQTState_video(QuickTimeGenericRTPSource::QTState const& qtState,
 		   unsigned& fourcc) {
--- a/libmpdemux/demux_rtp_internal.h	Tue Mar 11 19:03:31 2003 +0000
+++ b/libmpdemux/demux_rtp_internal.h	Tue Mar 11 19:08:31 2003 +0000
@@ -16,6 +16,10 @@
 #include <liveMedia.hh>
 #endif
 
+#if (LIVEMEDIA_LIBRARY_VERSION_INT < 1046649600)
+#error Please upgrade to version 2003.03.03 or later of the "LIVE.COM Streaming Media" libraries - available from <www.live.com/liveMedia/>
+#endif
+
 // Codec-specific initialization routines:
 void rtpCodecInitialize_video(demuxer_t* demuxer,
 			      MediaSubsession* subsession, unsigned& flags);
@@ -23,14 +27,19 @@
 			      MediaSubsession* subsession, unsigned& flags);
 
 // Flags that may be set by the above routines:
-#define RTPSTATE_IS_MPEG 0x1 // is an MPEG audio, video or transport stream
+#define RTPSTATE_IS_MPEG12_VIDEO 0x1 // is a MPEG-1 or 2 video stream
 
 // A routine to wait for the first packet of a RTP stream to arrive.
 // (For some RTP payload formats, codecs cannot be fully initialized until
 // we've started receiving data.)
-Boolean awaitRTPPacket(demuxer_t* demuxer, unsigned streamType,
-		       unsigned char*& packetData, unsigned& packetDataLen);
+Boolean awaitRTPPacket(demuxer_t* demuxer, demux_stream_t* ds,
+		       unsigned char*& packetData, unsigned& packetDataLen,
+		       float& pts);
     // "streamType": 0 => video; 1 => audio
     // This routine returns False if the input stream has closed
 
+// A routine for adding our own data to an incoming RTP data stream:
+Boolean insertRTPData(demuxer_t* demuxer, demux_stream_t* ds,
+		      unsigned char* data, unsigned dataLen);
+
 #endif