ios - AudioConverter#FillComplexBuffer 返回 -50 并且不转换任何内容

标签 ios xamarin.ios core-audio

我强烈关注 this Xamarin sample (基于 this Apple sample )将 LinearPCM 文件转换为 AAC 文件。

示例效果很好,但在我的项目中实现时,FillComplexBuffer 方法返回错误 -50 并且 InputData 事件不会触发一次, 因此没有任何转换。

只有在设备上测试时才会出现该错误。在模拟器上进行测试时,一切顺利,最后我得到了一个编码良好的 AAC 文件。

我今天尝试了很多东西,我没有发现我的代码和示例代码之间有任何区别。你知道这可能来自哪里吗?

我不知道这是否与 Xamarin 有任何关系,但似乎并非如此,因为 Xamarin 示例运行良好。

这是我的代码的相关部分:

protected void Encode(string path)
{
  // In class setup. File at TempWavFilePath has DecodedFormat as format.
  // 
  // DecodedFormat = AudioStreamBasicDescription.CreateLinearPCM();
  // AudioStreamBasicDescription encodedFormat = new AudioStreamBasicDescription()
  // {
  //   Format = AudioFormatType.MPEG4AAC,
  //   SampleRate = DecodedFormat.SampleRate,
  //   ChannelsPerFrame = DecodedFormat.ChannelsPerFrame,
  // };
  // AudioStreamBasicDescription.GetFormatInfo (ref encodedFormat);
  // EncodedFormat = encodedFormat;


  // Setup converter
  AudioStreamBasicDescription inputFormat = DecodedFormat;
  AudioStreamBasicDescription outputFormat = EncodedFormat;

  AudioConverterError converterCreateError;
  AudioConverter converter = AudioConverter.Create(inputFormat, outputFormat, out converterCreateError);
  if (converterCreateError != AudioConverterError.None)
  {
    Console.WriteLine("Converter creation error: " + converterCreateError);
  }
  converter.EncodeBitRate = 192000; // AAC 192kbps

  // get the actual formats back from the Audio Converter
  inputFormat = converter.CurrentInputStreamDescription;
  outputFormat = converter.CurrentOutputStreamDescription;


  /*** INPUT ***/

  AudioFile inputFile = AudioFile.OpenRead(NSUrl.FromFilename(TempWavFilePath));

  // init buffer
  const int inputBufferBytesSize = 32768;
  IntPtr inputBufferPtr = Marshal.AllocHGlobal(inputBufferBytesSize);

  // calc number of packets per read
  int inputSizePerPacket = inputFormat.BytesPerPacket;
  int inputBufferPacketSize = inputBufferBytesSize / inputSizePerPacket;
  AudioStreamPacketDescription[] inputPacketDescriptions = null;

  // init position
  long inputFilePosition = 0;

  // define input delegate
  converter.InputData += delegate(ref int numberDataPackets, AudioBuffers data, ref AudioStreamPacketDescription[] dataPacketDescription)
  {
    // how much to read
    if (numberDataPackets > inputBufferPacketSize)
    {
      numberDataPackets = inputBufferPacketSize;
    }

    // read from the file
    int outNumBytes;
    AudioFileError readError = inputFile.ReadPackets(false, out outNumBytes, inputPacketDescriptions, inputFilePosition, ref numberDataPackets, inputBufferPtr);
    if (readError != 0)
    {
      Console.WriteLine("Read error: " + readError);
    }

    // advance input file packet position
    inputFilePosition += numberDataPackets;

    // put the data pointer into the buffer list
    data.SetData(0, inputBufferPtr, outNumBytes);

    // add packet descriptions if required
    if (dataPacketDescription != null)
    {
      if (inputPacketDescriptions != null)
      {
        dataPacketDescription = inputPacketDescriptions;
      }
      else
      {
        dataPacketDescription = null;
      }
    }

    return AudioConverterError.None;
  };


  /*** OUTPUT ***/

  // create the destination file 
  var outputFile = AudioFile.Create (NSUrl.FromFilename(path), AudioFileType.M4A, outputFormat, AudioFileFlags.EraseFlags);

  // init buffer
  const int outputBufferBytesSize = 32768;
  IntPtr outputBufferPtr = Marshal.AllocHGlobal(outputBufferBytesSize);

  AudioBuffers buffers = new AudioBuffers(1);

  // calc number of packet per write
  int outputSizePerPacket = outputFormat.BytesPerPacket;
  AudioStreamPacketDescription[] outputPacketDescriptions = null;

  if (outputSizePerPacket == 0) {
    // if the destination format is VBR, we need to get max size per packet from the converter
    outputSizePerPacket = (int)converter.MaximumOutputPacketSize;

    // allocate memory for the PacketDescription structures describing the layout of each packet
    outputPacketDescriptions = new AudioStreamPacketDescription [outputBufferBytesSize / outputSizePerPacket];
  }
  int outputBufferPacketSize = outputBufferBytesSize / outputSizePerPacket;

  // init position
  long outputFilePosition = 0;

  long totalOutputFrames = 0; // used for debugging

  // write magic cookie if necessary
  if (converter.CompressionMagicCookie != null && converter.CompressionMagicCookie.Length != 0)
  {
    outputFile.MagicCookie = converter.CompressionMagicCookie;
  }

  // loop to convert data
  Console.WriteLine ("Converting...");
  while (true)
  {
    // create buffer
    buffers[0] = new AudioBuffer()
    {
      NumberChannels = outputFormat.ChannelsPerFrame,
      DataByteSize = outputBufferBytesSize,
      Data = outputBufferPtr
    };

    int writtenPackets = outputBufferPacketSize;

    // LET'S CONVERT (it's about time...)
    AudioConverterError converterFillError = converter.FillComplexBuffer(ref writtenPackets, buffers, outputPacketDescriptions);
    if (converterFillError != AudioConverterError.None)
    {
      Console.WriteLine("FillComplexBuffer error: " + converterFillError);
    }

    if (writtenPackets == 0) // EOF
    {
      break;
    }

    // write to output file
    int inNumBytes = buffers[0].DataByteSize;

    AudioFileError writeError = outputFile.WritePackets(false, inNumBytes, outputPacketDescriptions, outputFilePosition, ref writtenPackets, outputBufferPtr);
    if (writeError != 0)
    {
      Console.WriteLine("WritePackets error: {0}", writeError);
    }

    // advance output file packet position
    outputFilePosition += writtenPackets;

    if (FlowFormat.FramesPerPacket != 0) { 
      // the format has constant frames per packet
      totalOutputFrames += (writtenPackets * FlowFormat.FramesPerPacket);
    } else {
      // variable frames per packet require doing this for each packet (adding up the number of sample frames of data in each packet)
      for (var i = 0; i < writtenPackets; ++i)
      {
        totalOutputFrames += outputPacketDescriptions[i].VariableFramesInPacket;
      }
    }
  }

  // write out any of the leading and trailing frames for compressed formats only
  if (outputFormat.BitsPerChannel == 0)
  {
    Console.WriteLine("Total number of output frames counted: {0}", totalOutputFrames); 
    WritePacketTableInfo(converter, outputFile);
  }

  // write the cookie again - sometimes codecs will update cookies at the end of a conversion
  if (converter.CompressionMagicCookie != null && converter.CompressionMagicCookie.Length != 0)
  {
    outputFile.MagicCookie = converter.CompressionMagicCookie;
  }

  // Clean everything
  Marshal.FreeHGlobal(inputBufferPtr);
  Marshal.FreeHGlobal(outputBufferPtr);
  converter.Dispose();
  outputFile.Dispose();

  // Remove temp file
  File.Delete(TempWavFilePath);
}

我已经看到了 this SO question ,但不详细的 C++/Obj-C 相关答案似乎不适合我的问题。

谢谢!

最佳答案

我终于找到了解决方案!

我只需要在转换文件之前声明 AVAudioSession 类别。

AVAudioSession.SharedInstance().SetCategory(AVAudioSessionCategory.AudioProcessing);
AVAudioSession.SharedInstance().SetActive(true);

由于我还使用 AudioQueueRenderOffline,我必须将类别设置为 AVAudioSessionCategory.PlayAndRecord 所以离线渲染和音频转换工作。

关于ios - AudioConverter#FillComplexBuffer 返回 -50 并且不转换任何内容,我们在Stack Overflow上找到一个类似的问题: https://stackoverflow.com/questions/24143378/

相关文章:

ios - 在 iOS 中滚动 TableView 时保留 Tableview 单元格值

swift - Xamarin 中的 CGMutablePath

c# - Xamarin 动态加载本地化文件

ios - 使用 CAGradientLayer 作为后备存储在 UIView 上画线

ios - 如何在swift中制作系统字体 "Black Italic"?

c# - 在 MonoTouch 上使用 HTTPS 的 HttpListener

c - 音频样本生成器多线程 OSX

iphone - 减小转换后的音频文件的大小

cocoa - 类似于 NSSound 的框架,可以工作,但不需要处理陡峭的学习曲线

ios - Swift iOS Facebook SDK : How to share an animated GIF?