OpenCores
URL https://opencores.org/ocsvn/openrisc/openrisc/trunk

Subversion Repositories openrisc

[/] [openrisc/] [trunk/] [gnu-dev/] [or1k-gcc/] [libjava/] [classpath/] [native/] [jni/] [gstreamer-peer/] [gstreamer_io_peer.c] - Rev 774

Compare with Previous | Blame | View Log

/* gstreamer_io_peer.c -- Implements native methods for class
   GStreamerNativePeer
   Copyright (C) 2007 Free Software Foundation, Inc.
 
   This file is part of GNU Classpath.
 
   GNU Classpath is free software; you can redistribute it and/or modify
   it under the terms of the GNU General Public License as published by
   the Free Software Foundation; either version 2, or (at your option)
   any later version.
 
   GNU Classpath is distributed in the hope that it will be useful, but
   WITHOUT ANY WARRANTY; without even the implied warranty of
   MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the GNU
   General Public License for more details.
 
   You should have received a copy of the GNU General Public License
   along with GNU Classpath; see the file COPYING.  If not, write to the
   Free Software Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA
   02110-1301 USA.
 
   Linking this library statically or dynamically with other modules is
   making a combined work based on this library.  Thus, the terms and
   conditions of the GNU General Public License cover the whole
   combination.
 
   As a special exception, the copyright holders of this library give you
   permission to link this library with independent modules to produce an
   executable, regardless of the license terms of these independent
   modules, and to copy and distribute the resulting executable under
   terms of your choice, provided that you also meet, for each linked
   independent module, the terms and conditions of the license of that
   module.  An independent module is a module which is not derived from
   or based on this library.  If you modify this library, you may extend
   this exception to your version of the library, but you are not
   obligated to do so.  If you do not wish to do so, delete this
   exception statement from your version. */
 
#include <stdio.h>
#include <string.h>
 
#include <jni.h>
 
#include <glib.h>
#include <glib/gprintf.h>
 
#include <gdk/gdk.h>
 
#include <gst/gst.h>
 
#include "jcl.h"
 
#include "gst_peer.h"
 
#include "gnu_javax_sound_sampled_gstreamer_io_GstAudioFileReaderNativePeer.h"
 
#include "gst_classpath_src.h"
#include "gst_input_stream.h"
 
/* for caching */
static jfieldID fileFID = NULL;
static jfieldID pointerDataID = NULL;
 
static jfieldID mimetypeFID = NULL;
static jfieldID endiannessFID = NULL;
static jfieldID channelsFID = NULL;
static jfieldID rateFID = NULL;
static jfieldID widthFID = NULL;
static jfieldID depthFID = NULL;
static jfieldID isSignedFID = NULL;
static jfieldID nameFID = NULL;
static jfieldID layerFID = NULL;
static jfieldID bitrateFID = NULL;
static jfieldID framedFID = NULL;
static jfieldID typeFID = NULL;
 
typedef struct _AudioProperties AudioProperties;
struct _AudioProperties
{
  /*
   * NOTE: descriptions of the properties are taken from:
   * http://gstreamer.freedesktop.org/data/doc/gstreamer/head/pwg/html/section-types-definitions.html#table-audio-types
   */
 
  /* decoder name */
  const char *name;
 
  /* audio endiannes */
  const char *endianness;
 
  /* header size */
  const char *header_size;
 
  /* mime */  
  const char *mimetype;
 
  /* The sample rate of the data, in samples (per channel) per second */
  const char *samplerate;
 
  /* The number of channels of audio data */
  const char *channels;
 
  const char *layer;
 
  const char *bitrate;
 
  const char *framed;
 
  /*
   *  Defines if the values of the integer samples are signed or not.
   * Signed samples use one bit to indicate sign (negative or positive)
   * of the value. Unsigned samples are always positive.
   */
  const char *signess;
 
  /* */
  const char *rate;
 
  /* Number of bits allocated per sample. */
  const char *width;
 
  /*
   * The number of bits used per sample.
   * If the depth is less than the width, the low bits are assumed to be the
   * ones used. For example, a width of 32 and a depth of 24 means that
   * each sample is stored in a 32 bit word, but only the low
   * 24 bits are actually used.
   */
  const char *depth;
 
  /*
   * This is set in the case of the mpeg files.
   */
  const char *type;
 
  gboolean done;
};
 
/* ***** PRIVATE FUNCTIONS DECLARATION ***** */
 
static gboolean
set_strings (JNIEnv *env, const AudioProperties *properties, jobject header);
 
static gboolean
typefind_callback(GstElement *typefind, guint probability, const GstCaps *caps,
                  gpointer data);
 
static void
element_added (GstBin *bin, GstElement *element, gpointer data);
 
static void
new_decoded_pad (GstElement *decoder, GstPad *pad,
                 gboolean last, gpointer data);
 
static gboolean
fill_info (GstElement *decoder, AudioProperties *properties);
 
static gchar *
get_string_property (const GstStructure *structure, const gchar *property);
 
static gchar *
get_boolean_property (const GstStructure *structure, const gchar *property);
 
static gboolean
set_string (JNIEnv *env, jobject header, jfieldID fieldID,
            const gchar *property);
 
static void
free_properties (AudioProperties *properties);
 
static void
reset_properties (AudioProperties *properties);
 
static jboolean process_audio (GstElement *source, JNIEnv *env, jobject header);
 
/* ***** END: PRIVATE FUNCTIONS DECLARATION ***** */
 
/* ***** NATIVE FUNCTIONS ***** */
 
JNIEXPORT void JNICALL
Java_gnu_javax_sound_sampled_gstreamer_io_GstAudioFileReaderNativePeer_init_1id_1cache
  (JNIEnv *env, jclass clazz  __attribute__ ((unused)))
{
  jclass pointerClass = NULL;
  jclass GstHeader = NULL;
 
  GstHeader = JCL_FindClass(env, "gnu/javax/sound/sampled/gstreamer/io/GstAudioFileReaderNativePeer$GstHeader");
  fileFID = (*env)->GetFieldID(env, GstHeader, "file", "Ljava/lang/String;");
 
  mimetypeFID = (*env)->GetFieldID(env, GstHeader, "mimetype",
                                   "Ljava/lang/String;");
  endiannessFID = (*env)->GetFieldID(env, GstHeader, "endianness",
                                     "Ljava/lang/String;");
  channelsFID = (*env)->GetFieldID(env, GstHeader, "channels",
                                   "Ljava/lang/String;");
  rateFID = (*env)->GetFieldID(env, GstHeader, "rate", "Ljava/lang/String;");
  widthFID = (*env)->GetFieldID(env, GstHeader, "width", "Ljava/lang/String;");
  depthFID = (*env)->GetFieldID(env, GstHeader, "depth", "Ljava/lang/String;");
  isSignedFID = (*env)->GetFieldID(env, GstHeader, "isSigned",
                                   "Ljava/lang/String;");
  nameFID = (*env)->GetFieldID(env, GstHeader, "name", "Ljava/lang/String;");
  layerFID = (*env)->GetFieldID(env, GstHeader, "layer", "Ljava/lang/String;");
  bitrateFID = (*env)->GetFieldID(env, GstHeader, "bitrate",
                                  "Ljava/lang/String;");
  framedFID = (*env)->GetFieldID(env, GstHeader, "framed",
                                 "Ljava/lang/String;");
  typeFID = (*env)->GetFieldID(env, GstHeader, "type", "Ljava/lang/String;");
 
#if SIZEOF_VOID_P == 8
  pointerClass = JCL_FindClass (env, "gnu/classpath/Pointer64");
  if (pointerClass != NULL)
    {
      pointerDataID = (*env)->GetFieldID (env, pointerClass, "data", "J");
    }
#else
# if SIZEOF_VOID_P == 4
  pointerClass = JCL_FindClass (env, "gnu/classpath/Pointer32"); 
  if (pointerClass != NULL)
    { 
      pointerDataID = (*env)->GetFieldID(env, pointerClass, "data", "I");
    }
# else
#   error "Pointer size is not supported."
# endif /* SIZEOF_VOID_P == 4 */
#endif /* SIZEOF_VOID_P == 8 */
 
}
 
JNIEXPORT jboolean JNICALL
Java_gnu_javax_sound_sampled_gstreamer_io_GstAudioFileReaderNativePeer_gstreamer_1get_1audio_1format_1stream
  (JNIEnv *env, jclass clazz __attribute__ ((unused)), jobject header,
                                                       jobject pointer)
{
  GstInputStream *istream = NULL;
  GstElement *source = NULL;
  gboolean result = JNI_FALSE;
 
  if (header == NULL)
    return JNI_FALSE;
 
  if (pointer == NULL)
    return JNI_FALSE;
 
  gst_init (NULL, NULL);  
 
  istream = (GstInputStream *) get_object_from_pointer (env, pointer,
                                                        pointerDataID);
  if (istream == NULL)
    return JNI_FALSE;
 
  /* init gstreamer */
  gst_init (NULL, NULL);
 
  /* SOURCE */
  source = gst_element_factory_make ("classpathsrc", "source");
  if (source == NULL)
    {
      g_warning ("unable to create a source");
      return JNI_FALSE;
    }
  g_object_set (G_OBJECT (source), GST_CLASSPATH_SRC_ISTREAM, istream, NULL);
 
  result = process_audio (source, env, header);
 
  return result;
}
 
JNIEXPORT jboolean JNICALL
Java_gnu_javax_sound_sampled_gstreamer_io_GstAudioFileReaderNativePeer_gstreamer_1get_1audio_1format_1file
	(JNIEnv *env, jclass clazz __attribute__ ((unused)), jobject header)
{
  /* source file */
  const char *file = NULL;
 
  /* GStreamer elements */
  GstElement *source = NULL;
 
  jboolean result = JNI_FALSE;
 
  /* java fields */
  jstring _file = NULL;
 
  _file = (*env)->GetObjectField(env, header, fileFID);
  file = JCL_jstring_to_cstring (env, _file);
  if (file == NULL)
    {
      return JNI_FALSE;
    }
 
  gst_init (NULL, NULL);
 
  /* create the source element, will be used to read the file */
  source = gst_element_factory_make ("filesrc", "source");
  if (source == NULL)
    {
      JCL_free_cstring (env, _file, file);
      return JNI_FALSE;
    }
 
  /* set the file name */
  g_object_set (G_OBJECT (source), "location", file, NULL);
 
  result = process_audio (source, env, header);
 
  /* free stuff */
  JCL_free_cstring (env, _file, file);
 
  return result;
}
 
/* ***** END: NATIVE FUNCTIONS ***** */
 
/* ***** PRIVATE FUNCTIONS IMPLEMENTATION ***** */
 
static jboolean process_audio (GstElement *source, JNIEnv *env, jobject header)
{
  /* will contain the properties we need to put into the given GstHeader */
  AudioProperties *properties = NULL;
 
  /* GStreamer elements */
  GstElement *pipeline = NULL;
  GstElement *decoder = NULL;
 
  GstElement *typefind = NULL;
 
  GstStateChangeReturn res;
 
  jboolean result = JNI_FALSE;
 
  properties = (AudioProperties *) g_malloc0 (sizeof (AudioProperties));
  if (properties == NULL)
    {
      return result;
    }
  reset_properties(properties);
 
  /* 
   * create the decoder element, this will decode the stream and retrieve
   * its properties.
   * We connect a signal to this element, to be informed when it is done
   * in decoding the stream and to get the needed informations about the
   * audio file.
   */
  decoder = gst_element_factory_make ("decodebin", "decoder");
  if (decoder == NULL)
    {
      free_properties(properties);
      return result;
    }
 
  /* now, we create a pipeline and fill it with the other elements */
  pipeline = gst_pipeline_new ("pipeline");
  if (pipeline == NULL)
    {
      gst_object_unref (GST_OBJECT (decoder));
      free_properties(properties);   
      return result;
    }
 
  g_signal_connect (decoder, "new-decoded-pad", G_CALLBACK (new_decoded_pad),
                    pipeline);
  g_signal_connect (G_OBJECT (decoder), "element-added",
                    G_CALLBACK (element_added), properties);
 
  /*
   * we get the typefind from the decodebin to catch the additional properties
   * that the decodebin does not expose to us
   */
  typefind = gst_bin_get_by_name (GST_BIN (decoder), "typefind");
  if (typefind != NULL)
    {
      /* 
       * NOTE: the above is not a typo, we can live without the typefind,
       * just, our stream detection will not be as accurate as we would.
       * Anyway, if this fails, there is some problem, probabily a memory
       * error.
       */
       g_signal_connect (G_OBJECT (typefind), "have-type",
                         G_CALLBACK (typefind_callback), properties);
    }
 
  gst_bin_add_many (GST_BIN (pipeline), source, decoder, NULL);
  gst_element_link (source, decoder);
 
  /* 
   * now, we set the pipeline playing state to pause and traverse it
   * to get the info we need.
   */
 
  res = gst_element_set_state (pipeline, GST_STATE_PAUSED);
  if (res == GST_STATE_CHANGE_FAILURE)
    {
      gst_element_set_state (pipeline, GST_STATE_NULL);
      gst_object_unref (GST_OBJECT (pipeline));
 
      free_properties(properties);
 
      return result;
    }
 
  res = gst_element_get_state (pipeline, NULL, NULL, GST_CLOCK_TIME_NONE);
  if (res != GST_STATE_CHANGE_SUCCESS)
    {
      gst_element_set_state (pipeline, GST_STATE_NULL);
      gst_object_unref (GST_OBJECT (pipeline));
 
      free_properties(properties);
 
      return result;
    }
 
  if (fill_info (decoder, properties))
    {
      result = set_strings (env, properties, header);
    }
 
  /* free stuff */
  gst_element_set_state (pipeline, GST_STATE_NULL);
 
  free_properties (properties);
 
  gst_object_unref (GST_OBJECT (pipeline));
 
  return result;
}
 
 
static gboolean typefind_callback(GstElement *typefind __attribute__ ((unused)),
                                  guint probability __attribute__ ((unused)),
                                  const GstCaps *caps,
                                  gpointer data)
{
  GstStructure *structure = NULL;
  AudioProperties *properties = NULL;
 
  const char *mpeg = NULL;
 
  properties = (AudioProperties *) data;
 
  structure = gst_caps_get_structure (caps, 0);
 
  /* MIMETYPE */
  properties->mimetype = gst_structure_get_name (structure); 
  mpeg = get_string_property(structure, "mpegversion");
 
  if (mpeg != NULL)
    {
      properties->layer = get_string_property(structure, "layer");
      properties->type = (gchar *) g_malloc0 (_GST_MALLOC_SIZE_);
      g_snprintf ((gpointer) properties->type, _GST_MALLOC_SIZE_,
                  "MPEG%sV%s", mpeg,
                  properties->layer);
 
      g_free ((gpointer) mpeg);
    }
 
  return TRUE;
}
 
static void
new_decoded_pad (GstElement *decoder  __attribute__ ((unused)),
                 GstPad *pad,
                 gboolean last        __attribute__ ((unused)),
                 gpointer data)
{
  GstElement *pipeline = NULL;
  GstElement *fakesink = NULL;
  GstPad *sinkpad = NULL;
 
  pipeline = (GstElement *) data;
  if (pipeline == NULL)
    return;
 
  fakesink = gst_element_factory_make ("fakesink", NULL);
  if (fakesink == NULL)
    return;
 
  gst_bin_add (GST_BIN (pipeline), fakesink);
  sinkpad = gst_element_get_pad (fakesink, "sink");
  if (GST_PAD_LINK_FAILED (gst_pad_link (pad, sinkpad)))
    {
      gst_bin_remove (GST_BIN (pipeline), fakesink);
    }
  else
    {
      gst_element_set_state (fakesink, GST_STATE_PAUSED);
    }
}
 
static gboolean
set_strings (JNIEnv *env, const AudioProperties *properties, jobject header)
{
  gboolean result = FALSE;
 
  /* 
   * we only need at least one of them to be sure we can handle this
   * kind of audio data.
   */
 
  /* now, map our properties to the java class */
  set_string (env, header, mimetypeFID, properties->mimetype);
 
  if (set_string (env, header, endiannessFID, properties->endianness))
    result = JNI_TRUE;
 
  if (set_string (env, header, channelsFID, properties->channels))
    result = JNI_TRUE;
 
  if (set_string (env, header, rateFID, properties->rate))
    result = JNI_TRUE;
 
  if (set_string (env, header, widthFID, properties->width))
    result = JNI_TRUE;
 
  if (set_string (env, header, depthFID, properties->depth))
    result = JNI_TRUE;
 
  if (set_string (env, header, isSignedFID, properties->signess))
    result = JNI_TRUE;
 
  if (set_string (env, header, nameFID, properties->name))
    result = JNI_TRUE;
 
  /* non primary properties */
  set_string (env, header, layerFID, properties->layer);
  set_string (env, header, bitrateFID, properties->bitrate);
  set_string (env, header, framedFID, properties->framed);
  set_string (env, header, typeFID, properties->type);
 
  return result;    
}
 
static gboolean fill_info (GstElement *decoder, AudioProperties *properties)
{
  GstIterator *it = NULL;
  gpointer data = NULL;
  gboolean result = FALSE;
 
  it = gst_element_iterate_src_pads (decoder);
  while (gst_iterator_next (it, &data) == GST_ITERATOR_OK)
    {
      GstPad *pad = GST_PAD (data);
      GstCaps *caps;
 
      GstStructure *structure;
 
      const gchar *caps_string = NULL;
 
      caps = gst_pad_get_caps (pad);
      caps_string = gst_caps_to_string (caps);
 
      if (g_str_has_prefix (caps_string, "video"))
        {
          /* no video support, this is an audio library */
 
          g_free ((gpointer) caps_string);
          gst_caps_unref (caps);
          gst_object_unref (pad);
 
          continue; 
        }
 
      g_free ((gpointer) caps_string);
 
      structure = gst_caps_get_structure (GST_CAPS (caps), 0);
 
      /* fill the properties we need */
 
      /* SIGNESS */
      properties->signess = get_boolean_property(structure, "signed");
      if (properties->signess != NULL)
        {
          result = TRUE;
        }
 
      /* ENDIANNESS */
      properties->endianness = get_string_property(structure, "endianness");
      if (properties->endianness != NULL)
        {
          result = TRUE;
        }
 
      /* CHANNELS */
      properties->channels = get_string_property(structure, "channels");
      if (properties->channels != NULL)
        {
          result = TRUE;
        }
 
      /* RATE */
      properties->rate = get_string_property(structure, "rate");
      if (properties->rate != NULL)
        {
          result = TRUE;
        }
 
      /* WIDTH */
      properties->width = get_string_property(structure, "width");
      if (properties->width != NULL)
        {
          result = TRUE;
        }
 
      /* DEPTH */
      properties->depth = get_string_property(structure, "depth");
      if (properties->depth != NULL)
        {
          result = TRUE;
        }
 
      gst_caps_unref (caps);
      gst_object_unref (pad);
    }
 
    return result;
}
 
static void
free_properties (AudioProperties *properties __attribute__ ((unused)))
{
  /* FIXME this causes a segfault, a string not allocated by us? double free? */
  /*
  if (properties->name != NULL) g_free((gpointer) properties->name);
  if (properties->endianness != NULL) g_free((gpointer) properties->endianness);
  if (properties->channels != NULL) g_free((gpointer) properties->channels);
  if (properties->rate != NULL) g_free((gpointer) properties->rate);
  if (properties->width != NULL) g_free((gpointer) properties->width);
  if (properties->depth != NULL) g_free((gpointer) properties->depth);
  if (properties->layer != NULL) g_free((gpointer) properties->layer);
  if (properties->bitrate != NULL) g_free((gpointer) properties->bitrate);
  if (properties->framed != NULL) g_free((gpointer) properties->framed);
 
  if (properties != NULL) g_free ((gpointer) properties);
  */
}
 
static void reset_properties (AudioProperties *properties)
{
  properties->done = FALSE;
  properties->signess = FALSE;
  properties->name = NULL;
  properties->endianness = NULL;
  properties->channels = NULL;
  properties->rate = NULL;
  properties->width  = NULL;
  properties->depth = NULL;
  properties->layer = NULL;
  properties->bitrate = NULL;
  properties->framed = NULL;
}
 
static gchar *get_string_property (const GstStructure *structure,
                                   const gchar *property)
{
  int props = 0;
  gchar *result = NULL;
 
  if (property == NULL)
    {
      return NULL;
    }
 
  /* we don't need more */
  result = (gchar *) g_malloc0 (_GST_MALLOC_SIZE_);
  if (result == NULL)
    {
      /* huston, we have a problem here... */
      return NULL;
    }
 
  if (gst_structure_get_int (structure, property, &props))
    {
      g_snprintf (result, _GST_MALLOC_SIZE_, "%d", props);
    }
  else
    {
      g_free ((gpointer) result);
      return NULL;
    }
 
  return result;
}
 
static gchar *get_boolean_property (const GstStructure *structure,
                                    const gchar *property)
{
  gchar *result = NULL;
  gboolean props = FALSE;
 
  result = (gchar *) g_malloc0 (_GST_MALLOC_SIZE_);
  if (result == NULL)
    {
      /* huston, we have a problem here... */
      return NULL;
    }
 
  if (gst_structure_get_boolean (structure, property, &props))
    {
      g_snprintf (result, _GST_MALLOC_SIZE_, "%s", (props ? "true" : "false" ));
    }
  else
    {
      g_free ((gpointer) result);
      return NULL;
    }
 
  return result;
}
 
static gboolean set_string (JNIEnv *env, jobject header, jfieldID fieldID,
                            const gchar *property)
{ 
  jstring property_string_field = NULL; 
 
  if (property == NULL || header == NULL)
    {
      return JNI_FALSE;
    }
 
  property_string_field = (*env)->NewStringUTF(env, property);
  if (property_string_field == NULL)
    {
      return JNI_FALSE;
    }
 
  (*env)->SetObjectField(env, header, fieldID, property_string_field);
 
  return JNI_TRUE;
}
 
static void element_added (GstBin *bin, GstElement *element, gpointer data)
{
  GstElementFactory *factory;
 
  factory = gst_element_get_factory (element);
  ((AudioProperties *) data)->name = gst_element_factory_get_longname (factory);
}
 
/* ***** END: PRIVATE FUNCTIONS IMPLEMENTATION ***** */
 

Compare with Previous | Blame | View Log

powered by: WebSVN 2.1.0

© copyright 1999-2024 OpenCores.org, equivalent to Oliscience, all rights reserved. OpenCores®, registered trademark.