Update web implementation to use platform interface.
This commit is contained in:
parent
d07d0f358a
commit
1be1c212f6
11 changed files with 323 additions and 308 deletions
|
@ -6,7 +6,7 @@
|
|||
import 'dart:ui';
|
||||
|
||||
import 'package:audio_session/audio_session_web.dart';
|
||||
import 'package:just_audio/just_audio_web.dart';
|
||||
import 'package:just_audio_web/just_audio_web.dart';
|
||||
|
||||
import 'package:flutter_web_plugins/flutter_web_plugins.dart';
|
||||
|
||||
|
|
|
@ -121,6 +121,13 @@ packages:
|
|||
relative: true
|
||||
source: path
|
||||
version: "0.0.1"
|
||||
just_audio_web:
|
||||
dependency: "direct dev"
|
||||
description:
|
||||
path: "../../just_audio_web"
|
||||
relative: true
|
||||
source: path
|
||||
version: "0.0.1"
|
||||
matcher:
|
||||
dependency: transitive
|
||||
description:
|
||||
|
|
|
@ -18,6 +18,8 @@ dev_dependencies:
|
|||
|
||||
just_audio:
|
||||
path: ../
|
||||
just_audio_web:
|
||||
path: ../../just_audio_web
|
||||
|
||||
# For information on the generic Dart part of this file, see the
|
||||
# following page: https://dart.dev/tools/pub/pubspec
|
||||
|
|
|
@ -1,7 +1,7 @@
|
|||
name: just_audio
|
||||
description: Flutter plugin to play audio from streams, files, assets, DASH/HLS streams and playlists. Works with audio_service to play audio in the background.
|
||||
version: 0.4.5
|
||||
homepage: https://github.com/ryanheise/just_audio
|
||||
homepage: https://github.com/ryanheise/just_audio/tree/master/just_audio
|
||||
|
||||
environment:
|
||||
sdk: ">=2.7.0 <3.0.0"
|
||||
|
@ -35,6 +35,8 @@ flutter:
|
|||
pluginClass: JustAudioPlugin
|
||||
macos:
|
||||
pluginClass: JustAudioPlugin
|
||||
web:
|
||||
pluginClass: JustAudioPlugin
|
||||
fileName: just_audio_web.dart
|
||||
# web:
|
||||
# default_package: just_audio_web
|
||||
# web:
|
||||
# pluginClass: JustAudioPlugin
|
||||
# fileName: just_audio_web.dart
|
||||
|
|
|
@ -1,6 +1,6 @@
|
|||
name: just_audio_platform_interface
|
||||
description: A common platform interface for the just_audio plugin.
|
||||
homepage: https://github.com/ryanheise/just_audio/just_audio_platform_interface
|
||||
homepage: https://github.com/ryanheise/just_audio/tree/master/just_audio_platform_interface
|
||||
# NOTE: We strongly prefer non-breaking changes, even at the expense of a
|
||||
# less-clean API. See https://flutter.dev/go/platform-interface-breaking-changes
|
||||
version: 0.0.1
|
||||
|
|
44
just_audio_web/.gitignore
vendored
Normal file
44
just_audio_web/.gitignore
vendored
Normal file
|
@ -0,0 +1,44 @@
|
|||
# Miscellaneous
|
||||
*.class
|
||||
*.log
|
||||
*.pyc
|
||||
*.swp
|
||||
.DS_Store
|
||||
.atom/
|
||||
.buildlog/
|
||||
.history
|
||||
.svn/
|
||||
|
||||
# IntelliJ related
|
||||
*.iml
|
||||
*.ipr
|
||||
*.iws
|
||||
.idea/
|
||||
|
||||
# The .vscode folder contains launch configuration and tasks you configure in
|
||||
# VS Code which you may wish to be included in version control, so this line
|
||||
# is commented out by default.
|
||||
#.vscode/
|
||||
|
||||
# Flutter/Dart/Pub related
|
||||
**/doc/api/
|
||||
**/ios/Flutter/.last_build_id
|
||||
.dart_tool/
|
||||
.flutter-plugins
|
||||
.flutter-plugins-dependencies
|
||||
.packages
|
||||
.pub-cache/
|
||||
.pub/
|
||||
/build/
|
||||
|
||||
# Web related
|
||||
lib/generated_plugin_registrant.dart
|
||||
|
||||
# Symbolication related
|
||||
app.*.symbols
|
||||
|
||||
# Obfuscation related
|
||||
app.*.map.json
|
||||
|
||||
# Exceptions to above rules.
|
||||
!/packages/flutter_tools/test/data/dart_dependencies_test/**/.packages
|
10
just_audio_web/.metadata
Normal file
10
just_audio_web/.metadata
Normal file
|
@ -0,0 +1,10 @@
|
|||
# This file tracks properties of this Flutter project.
|
||||
# Used by Flutter tool to assess capabilities and perform upgrades etc.
|
||||
#
|
||||
# This file should be version controlled and should not be manually edited.
|
||||
|
||||
version:
|
||||
revision: bbfbf1770cca2da7c82e887e4e4af910034800b6
|
||||
channel: stable
|
||||
|
||||
project_type: app
|
5
just_audio_web/README.md
Normal file
5
just_audio_web/README.md
Normal file
|
@ -0,0 +1,5 @@
|
|||
# just_audio_web
|
||||
|
||||
The web implementation of [`just_audio`][1].
|
||||
|
||||
[1]: ../just_audio
|
|
@ -5,127 +5,33 @@ import 'dart:math';
|
|||
import 'package:flutter/services.dart';
|
||||
import 'package:flutter/widgets.dart';
|
||||
import 'package:flutter_web_plugins/flutter_web_plugins.dart';
|
||||
import 'package:just_audio/just_audio.dart';
|
||||
import 'package:just_audio_platform_interface/just_audio_platform_interface.dart';
|
||||
|
||||
final Random _random = Random();
|
||||
|
||||
class JustAudioPlugin {
|
||||
class JustAudioPlugin extends JustAudioPlatform {
|
||||
static void registerWith(Registrar registrar) {
|
||||
final MethodChannel channel = MethodChannel(
|
||||
'com.ryanheise.just_audio.methods',
|
||||
const StandardMethodCodec(),
|
||||
registrar.messenger);
|
||||
final JustAudioPlugin instance = JustAudioPlugin(registrar);
|
||||
channel.setMethodCallHandler(instance.handleMethodCall);
|
||||
JustAudioPlatform.instance = JustAudioPlugin();
|
||||
}
|
||||
|
||||
final Registrar registrar;
|
||||
|
||||
JustAudioPlugin(this.registrar);
|
||||
|
||||
Future<dynamic> handleMethodCall(MethodCall call) async {
|
||||
switch (call.method) {
|
||||
case 'init':
|
||||
final String id = call.arguments[0];
|
||||
new Html5AudioPlayer(id: id, registrar: registrar);
|
||||
return null;
|
||||
case 'setIosCategory':
|
||||
return null;
|
||||
default:
|
||||
throw PlatformException(code: 'Unimplemented');
|
||||
}
|
||||
Future<AudioPlayerPlatform> init(InitRequest request) async {
|
||||
return Html5AudioPlayer(id: request.id);
|
||||
}
|
||||
}
|
||||
|
||||
abstract class JustAudioPlayer {
|
||||
abstract class JustAudioPlayer extends AudioPlayerPlatform {
|
||||
final String id;
|
||||
final Registrar registrar;
|
||||
final MethodChannel methodChannel;
|
||||
final PluginEventChannel eventChannel;
|
||||
final StreamController eventController = StreamController();
|
||||
ProcessingState _processingState = ProcessingState.none;
|
||||
final eventController = StreamController<PlaybackEventMessage>();
|
||||
ProcessingStateMessage _processingState = ProcessingStateMessage.none;
|
||||
bool _playing = false;
|
||||
int _index;
|
||||
|
||||
JustAudioPlayer({@required this.id, @required this.registrar})
|
||||
: methodChannel = MethodChannel('com.ryanheise.just_audio.methods.$id',
|
||||
const StandardMethodCodec(), registrar.messenger),
|
||||
eventChannel = PluginEventChannel('com.ryanheise.just_audio.events.$id',
|
||||
const StandardMethodCodec(), registrar.messenger) {
|
||||
methodChannel.setMethodCallHandler(_methodHandler);
|
||||
eventChannel.controller = eventController;
|
||||
}
|
||||
|
||||
Future<dynamic> _methodHandler(MethodCall call) async {
|
||||
try {
|
||||
final args = call.arguments;
|
||||
switch (call.method) {
|
||||
case 'load':
|
||||
return await load(args[0]);
|
||||
case 'play':
|
||||
return await play();
|
||||
case 'pause':
|
||||
return await pause();
|
||||
case 'setVolume':
|
||||
return await setVolume(args[0]);
|
||||
case 'setSpeed':
|
||||
return await setSpeed(args[0]);
|
||||
case 'setLoopMode':
|
||||
return await setLoopMode(args[0]);
|
||||
case 'setShuffleModeEnabled':
|
||||
return await setShuffleModeEnabled(args[0]);
|
||||
case 'setAutomaticallyWaitsToMinimizeStalling':
|
||||
return null;
|
||||
case 'seek':
|
||||
return await seek(args[0], args[1]);
|
||||
case 'dispose':
|
||||
return dispose();
|
||||
case 'concatenating.add':
|
||||
return await concatenatingAdd(args[0], args[1]);
|
||||
case "concatenating.insert":
|
||||
return await concatenatingInsert(args[0], args[1], args[2]);
|
||||
case "concatenating.addAll":
|
||||
return await concatenatingAddAll(args[0], args[1]);
|
||||
case "concatenating.insertAll":
|
||||
return await concatenatingInsertAll(args[0], args[1], args[2]);
|
||||
case "concatenating.removeAt":
|
||||
return await concatenatingRemoveAt(args[0], args[1]);
|
||||
case "concatenating.removeRange":
|
||||
return await concatenatingRemoveRange(args[0], args[1], args[2]);
|
||||
case "concatenating.move":
|
||||
return await concatenatingMove(args[0], args[1], args[2]);
|
||||
case "concatenating.clear":
|
||||
return await concatenatingClear(args[0]);
|
||||
case "setAndroidAudioAttributes":
|
||||
return null;
|
||||
default:
|
||||
throw PlatformException(code: 'Unimplemented');
|
||||
}
|
||||
} catch (e, stacktrace) {
|
||||
print("$stacktrace");
|
||||
rethrow;
|
||||
}
|
||||
}
|
||||
|
||||
Future<int> load(Map source);
|
||||
|
||||
Future<void> play();
|
||||
|
||||
Future<void> pause();
|
||||
|
||||
Future<void> setVolume(double volume);
|
||||
|
||||
Future<void> setSpeed(double speed);
|
||||
|
||||
Future<void> setLoopMode(int mode);
|
||||
|
||||
Future<void> setShuffleModeEnabled(bool enabled);
|
||||
|
||||
Future<void> seek(int position, int index);
|
||||
JustAudioPlayer({@required this.id});
|
||||
|
||||
@mustCallSuper
|
||||
void dispose() {
|
||||
Future<DisposeResponse> dispose(DisposeRequest request) async {
|
||||
eventController.close();
|
||||
return DisposeResponse();
|
||||
}
|
||||
|
||||
Duration getCurrentPosition();
|
||||
|
@ -134,37 +40,22 @@ abstract class JustAudioPlayer {
|
|||
|
||||
Duration getDuration();
|
||||
|
||||
concatenatingAdd(String playerId, Map source);
|
||||
|
||||
concatenatingInsert(String playerId, int index, Map source);
|
||||
|
||||
concatenatingAddAll(String playerId, List sources);
|
||||
|
||||
concatenatingInsertAll(String playerId, int index, List sources);
|
||||
|
||||
concatenatingRemoveAt(String playerId, int index);
|
||||
|
||||
concatenatingRemoveRange(String playerId, int start, int end);
|
||||
|
||||
concatenatingMove(String playerId, int currentIndex, int newIndex);
|
||||
|
||||
concatenatingClear(String playerId);
|
||||
|
||||
broadcastPlaybackEvent() {
|
||||
var updateTime = DateTime.now().millisecondsSinceEpoch;
|
||||
eventController.add({
|
||||
'processingState': _processingState.index,
|
||||
'updatePosition': getCurrentPosition()?.inMilliseconds,
|
||||
'updateTime': updateTime,
|
||||
'bufferedPosition': getBufferedPosition()?.inMilliseconds,
|
||||
var updateTime = DateTime.now();
|
||||
eventController.add(PlaybackEventMessage(
|
||||
processingState: _processingState,
|
||||
updatePosition: getCurrentPosition(),
|
||||
updateTime: updateTime,
|
||||
bufferedPosition: getBufferedPosition(),
|
||||
// TODO: Icy Metadata
|
||||
'icyMetadata': null,
|
||||
'duration': getDuration()?.inMilliseconds,
|
||||
'currentIndex': _index,
|
||||
});
|
||||
icyMetadata: null,
|
||||
duration: getDuration(),
|
||||
currentIndex: _index,
|
||||
androidAudioSessionId: null,
|
||||
));
|
||||
}
|
||||
|
||||
transition(ProcessingState processingState) {
|
||||
transition(ProcessingStateMessage processingState) {
|
||||
_processingState = processingState;
|
||||
broadcastPlaybackEvent();
|
||||
}
|
||||
|
@ -174,12 +65,11 @@ class Html5AudioPlayer extends JustAudioPlayer {
|
|||
AudioElement _audioElement = AudioElement();
|
||||
Completer _durationCompleter;
|
||||
AudioSourcePlayer _audioSourcePlayer;
|
||||
LoopMode _loopMode = LoopMode.off;
|
||||
LoopModeMessage _loopMode = LoopModeMessage.off;
|
||||
bool _shuffleModeEnabled = false;
|
||||
final Map<String, AudioSourcePlayer> _audioSourcePlayers = {};
|
||||
|
||||
Html5AudioPlayer({@required String id, @required Registrar registrar})
|
||||
: super(id: id, registrar: registrar) {
|
||||
Html5AudioPlayer({@required String id}) : super(id: id) {
|
||||
_audioElement.addEventListener('durationchange', (event) {
|
||||
_durationCompleter?.complete();
|
||||
broadcastPlaybackEvent();
|
||||
|
@ -194,16 +84,16 @@ class Html5AudioPlayer extends JustAudioPlayer {
|
|||
_currentAudioSourcePlayer.timeUpdated(_audioElement.currentTime);
|
||||
});
|
||||
_audioElement.addEventListener('loadstart', (event) {
|
||||
transition(ProcessingState.buffering);
|
||||
transition(ProcessingStateMessage.buffering);
|
||||
});
|
||||
_audioElement.addEventListener('waiting', (event) {
|
||||
transition(ProcessingState.buffering);
|
||||
transition(ProcessingStateMessage.buffering);
|
||||
});
|
||||
_audioElement.addEventListener('stalled', (event) {
|
||||
transition(ProcessingState.buffering);
|
||||
transition(ProcessingStateMessage.buffering);
|
||||
});
|
||||
_audioElement.addEventListener('canplaythrough', (event) {
|
||||
transition(ProcessingState.ready);
|
||||
transition(ProcessingStateMessage.ready);
|
||||
});
|
||||
_audioElement.addEventListener('progress', (event) {
|
||||
broadcastPlaybackEvent();
|
||||
|
@ -232,9 +122,9 @@ class Html5AudioPlayer extends JustAudioPlayer {
|
|||
}
|
||||
|
||||
onEnded() async {
|
||||
if (_loopMode == LoopMode.one) {
|
||||
await seek(0, null);
|
||||
play();
|
||||
if (_loopMode == LoopModeMessage.one) {
|
||||
await _seek(0, null);
|
||||
_play();
|
||||
} else {
|
||||
final order = this.order;
|
||||
final orderInv = getInv(order);
|
||||
|
@ -244,25 +134,25 @@ class Html5AudioPlayer extends JustAudioPlayer {
|
|||
await _currentAudioSourcePlayer.load();
|
||||
// Should always be true...
|
||||
if (_playing) {
|
||||
play();
|
||||
_play();
|
||||
}
|
||||
} else {
|
||||
// reached end of playlist
|
||||
if (_loopMode == LoopMode.all) {
|
||||
if (_loopMode == LoopModeMessage.all) {
|
||||
// Loop back to the beginning
|
||||
if (order.length == 1) {
|
||||
await seek(0, null);
|
||||
play();
|
||||
await _seek(0, null);
|
||||
_play();
|
||||
} else {
|
||||
_index = order[0];
|
||||
await _currentAudioSourcePlayer.load();
|
||||
// Should always be true...
|
||||
if (_playing) {
|
||||
play();
|
||||
_play();
|
||||
}
|
||||
}
|
||||
} else {
|
||||
transition(ProcessingState.completed);
|
||||
transition(ProcessingStateMessage.completed);
|
||||
}
|
||||
}
|
||||
}
|
||||
|
@ -275,18 +165,22 @@ class Html5AudioPlayer extends JustAudioPlayer {
|
|||
: null;
|
||||
|
||||
@override
|
||||
Future<int> load(Map source) async {
|
||||
Stream<PlaybackEventMessage> get playbackEventMessageStream =>
|
||||
eventController.stream;
|
||||
|
||||
@override
|
||||
Future<LoadResponse> load(LoadRequest request) async {
|
||||
_currentAudioSourcePlayer?.pause();
|
||||
_audioSourcePlayer = getAudioSource(source);
|
||||
_audioSourcePlayer = getAudioSource(request.audioSourceMessage);
|
||||
_index = 0;
|
||||
if (_shuffleModeEnabled) {
|
||||
_audioSourcePlayer?.shuffle(0, _index);
|
||||
}
|
||||
return (await _currentAudioSourcePlayer.load())?.inMilliseconds;
|
||||
return LoadResponse(duration: await _currentAudioSourcePlayer.load());
|
||||
}
|
||||
|
||||
Future<Duration> loadUri(final Uri uri) async {
|
||||
transition(ProcessingState.loading);
|
||||
transition(ProcessingStateMessage.loading);
|
||||
final src = uri.toString();
|
||||
if (src != _audioElement.src) {
|
||||
_durationCompleter = Completer<num>();
|
||||
|
@ -302,7 +196,7 @@ class Html5AudioPlayer extends JustAudioPlayer {
|
|||
_durationCompleter = null;
|
||||
}
|
||||
}
|
||||
transition(ProcessingState.ready);
|
||||
transition(ProcessingStateMessage.ready);
|
||||
final seconds = _audioElement.duration;
|
||||
return seconds.isFinite
|
||||
? Duration(milliseconds: (seconds * 1000).toInt())
|
||||
|
@ -310,42 +204,58 @@ class Html5AudioPlayer extends JustAudioPlayer {
|
|||
}
|
||||
|
||||
@override
|
||||
Future<void> play() async {
|
||||
Future<PlayResponse> play(PlayRequest request) async {
|
||||
await _play();
|
||||
return PlayResponse();
|
||||
}
|
||||
|
||||
Future<void> _play() async {
|
||||
_playing = true;
|
||||
await _currentAudioSourcePlayer.play();
|
||||
}
|
||||
|
||||
@override
|
||||
Future<void> pause() async {
|
||||
Future<PauseResponse> pause(PauseRequest request) async {
|
||||
_playing = false;
|
||||
_currentAudioSourcePlayer.pause();
|
||||
return PauseResponse();
|
||||
}
|
||||
|
||||
@override
|
||||
Future<void> setVolume(double volume) async {
|
||||
_audioElement.volume = volume;
|
||||
Future<SetVolumeResponse> setVolume(SetVolumeRequest request) async {
|
||||
_audioElement.volume = request.volume;
|
||||
return SetVolumeResponse();
|
||||
}
|
||||
|
||||
@override
|
||||
Future<void> setSpeed(double speed) async {
|
||||
_audioElement.playbackRate = speed;
|
||||
Future<SetSpeedResponse> setSpeed(SetSpeedRequest request) async {
|
||||
_audioElement.playbackRate = request.speed;
|
||||
return SetSpeedResponse();
|
||||
}
|
||||
|
||||
@override
|
||||
Future<void> setLoopMode(int mode) async {
|
||||
_loopMode = LoopMode.values[mode];
|
||||
Future<SetLoopModeResponse> setLoopMode(SetLoopModeRequest request) async {
|
||||
_loopMode = request.loopMode;
|
||||
return SetLoopModeResponse();
|
||||
}
|
||||
|
||||
@override
|
||||
Future<void> setShuffleModeEnabled(bool enabled) async {
|
||||
_shuffleModeEnabled = enabled;
|
||||
if (enabled) {
|
||||
Future<SetShuffleModeResponse> setShuffleMode(
|
||||
SetShuffleModeRequest request) async {
|
||||
_shuffleModeEnabled = request.shuffleMode == ShuffleModeMessage.all;
|
||||
if (_shuffleModeEnabled) {
|
||||
_audioSourcePlayer?.shuffle(0, _index);
|
||||
}
|
||||
return SetShuffleModeResponse();
|
||||
}
|
||||
|
||||
@override
|
||||
Future<void> seek(int position, int newIndex) async {
|
||||
Future<SeekResponse> seek(SeekRequest request) async {
|
||||
await _seek(request.position.inMilliseconds, request.index);
|
||||
return SeekResponse();
|
||||
}
|
||||
|
||||
Future<void> _seek(int position, int newIndex) async {
|
||||
int index = newIndex ?? _index;
|
||||
if (index != _index) {
|
||||
_currentAudioSourcePlayer.pause();
|
||||
|
@ -363,89 +273,70 @@ class Html5AudioPlayer extends JustAudioPlayer {
|
|||
ConcatenatingAudioSourcePlayer _concatenating(String playerId) =>
|
||||
_audioSourcePlayers[playerId] as ConcatenatingAudioSourcePlayer;
|
||||
|
||||
concatenatingAdd(String playerId, Map source) {
|
||||
final playlist = _concatenating(playerId);
|
||||
playlist.add(getAudioSource(source));
|
||||
}
|
||||
|
||||
concatenatingInsert(String playerId, int index, Map source) {
|
||||
_concatenating(playerId).insert(index, getAudioSource(source));
|
||||
if (index <= _index) {
|
||||
_index++;
|
||||
@override
|
||||
Future<ConcatenatingInsertAllResponse> concatenatingInsertAll(
|
||||
ConcatenatingInsertAllRequest request) async {
|
||||
_concatenating(request.id)
|
||||
.insertAll(request.index, getAudioSources(request.children));
|
||||
if (request.index <= _index) {
|
||||
_index += request.children.length;
|
||||
}
|
||||
return ConcatenatingInsertAllResponse();
|
||||
}
|
||||
|
||||
concatenatingAddAll(String playerId, List sources) {
|
||||
_concatenating(playerId).addAll(getAudioSources(sources));
|
||||
}
|
||||
|
||||
concatenatingInsertAll(String playerId, int index, List sources) {
|
||||
_concatenating(playerId).insertAll(index, getAudioSources(sources));
|
||||
if (index <= _index) {
|
||||
_index += sources.length;
|
||||
}
|
||||
}
|
||||
|
||||
concatenatingRemoveAt(String playerId, int index) async {
|
||||
// Pause if removing current item
|
||||
if (_index == index && _playing) {
|
||||
_currentAudioSourcePlayer.pause();
|
||||
}
|
||||
_concatenating(playerId).removeAt(index);
|
||||
if (_index == index) {
|
||||
// Skip backward if there's nothing after this
|
||||
if (index == _audioSourcePlayer.sequence.length) {
|
||||
_index--;
|
||||
}
|
||||
// Resume playback at the new item (if it exists)
|
||||
if (_playing && _currentAudioSourcePlayer != null) {
|
||||
await _currentAudioSourcePlayer.load();
|
||||
_currentAudioSourcePlayer.play();
|
||||
}
|
||||
} else if (index < _index) {
|
||||
// Reflect that the current item has shifted its position
|
||||
_index--;
|
||||
}
|
||||
}
|
||||
|
||||
concatenatingRemoveRange(String playerId, int start, int end) async {
|
||||
if (_index >= start && _index < end && _playing) {
|
||||
@override
|
||||
Future<ConcatenatingRemoveRangeResponse> concatenatingRemoveRange(
|
||||
ConcatenatingRemoveRangeRequest request) async {
|
||||
if (_index >= request.startIndex && _index < request.endIndex && _playing) {
|
||||
// Pause if removing current item
|
||||
_currentAudioSourcePlayer.pause();
|
||||
}
|
||||
_concatenating(playerId).removeRange(start, end);
|
||||
if (_index >= start && _index < end) {
|
||||
_concatenating(request.id)
|
||||
.removeRange(request.startIndex, request.endIndex);
|
||||
if (_index >= request.startIndex && _index < request.endIndex) {
|
||||
// Skip backward if there's nothing after this
|
||||
if (start >= _audioSourcePlayer.sequence.length) {
|
||||
_index = start - 1;
|
||||
if (request.startIndex >= _audioSourcePlayer.sequence.length) {
|
||||
_index = request.startIndex - 1;
|
||||
} else {
|
||||
_index = start;
|
||||
_index = request.startIndex;
|
||||
}
|
||||
// Resume playback at the new item (if it exists)
|
||||
if (_playing && _currentAudioSourcePlayer != null) {
|
||||
await _currentAudioSourcePlayer.load();
|
||||
_currentAudioSourcePlayer.play();
|
||||
}
|
||||
} else if (end <= _index) {
|
||||
} else if (request.endIndex <= _index) {
|
||||
// Reflect that the current item has shifted its position
|
||||
_index -= (end - start);
|
||||
_index -= (request.endIndex - request.startIndex);
|
||||
}
|
||||
return ConcatenatingRemoveRangeResponse();
|
||||
}
|
||||
|
||||
concatenatingMove(String playerId, int currentIndex, int newIndex) {
|
||||
_concatenating(playerId).move(currentIndex, newIndex);
|
||||
if (currentIndex == _index) {
|
||||
_index = newIndex;
|
||||
} else if (currentIndex < _index && newIndex >= _index) {
|
||||
@override
|
||||
Future<ConcatenatingMoveResponse> concatenatingMove(
|
||||
ConcatenatingMoveRequest request) async {
|
||||
_concatenating(request.id).move(request.currentIndex, request.newIndex);
|
||||
if (request.currentIndex == _index) {
|
||||
_index = request.newIndex;
|
||||
} else if (request.currentIndex < _index && request.newIndex >= _index) {
|
||||
_index--;
|
||||
} else if (currentIndex > _index && newIndex <= _index) {
|
||||
} else if (request.currentIndex > _index && request.newIndex <= _index) {
|
||||
_index++;
|
||||
}
|
||||
return ConcatenatingMoveResponse();
|
||||
}
|
||||
|
||||
concatenatingClear(String playerId) {
|
||||
_currentAudioSourcePlayer.pause();
|
||||
_concatenating(playerId).clear();
|
||||
@override
|
||||
Future<SetAndroidAudioAttributesResponse> setAndroidAudioAttributes(
|
||||
SetAndroidAudioAttributesRequest request) async {
|
||||
return SetAndroidAudioAttributesResponse();
|
||||
}
|
||||
|
||||
@override
|
||||
Future<SetAutomaticallyWaitsToMinimizeStallingResponse>
|
||||
setAutomaticallyWaitsToMinimizeStalling(
|
||||
SetAutomaticallyWaitsToMinimizeStallingRequest request) async {
|
||||
return SetAutomaticallyWaitsToMinimizeStallingResponse();
|
||||
}
|
||||
|
||||
@override
|
||||
|
@ -458,57 +349,56 @@ class Html5AudioPlayer extends JustAudioPlayer {
|
|||
Duration getDuration() => _currentAudioSourcePlayer?.duration;
|
||||
|
||||
@override
|
||||
void dispose() {
|
||||
Future<DisposeResponse> dispose(DisposeRequest request) async {
|
||||
_currentAudioSourcePlayer?.pause();
|
||||
_audioElement.removeAttribute('src');
|
||||
_audioElement.load();
|
||||
transition(ProcessingState.none);
|
||||
super.dispose();
|
||||
transition(ProcessingStateMessage.none);
|
||||
return await super.dispose(request);
|
||||
}
|
||||
|
||||
List<AudioSourcePlayer> getAudioSources(List json) =>
|
||||
json.map((s) => getAudioSource(s)).toList();
|
||||
List<AudioSourcePlayer> getAudioSources(List messages) =>
|
||||
messages.map((message) => getAudioSource(message)).toList();
|
||||
|
||||
AudioSourcePlayer getAudioSource(Map json) {
|
||||
final String id = json['id'];
|
||||
AudioSourcePlayer getAudioSource(AudioSourceMessage audioSourceMessage) {
|
||||
final String id = audioSourceMessage.id;
|
||||
var audioSourcePlayer = _audioSourcePlayers[id];
|
||||
if (audioSourcePlayer == null) {
|
||||
audioSourcePlayer = decodeAudioSource(json);
|
||||
audioSourcePlayer = decodeAudioSource(audioSourceMessage);
|
||||
_audioSourcePlayers[id] = audioSourcePlayer;
|
||||
}
|
||||
return audioSourcePlayer;
|
||||
}
|
||||
|
||||
AudioSourcePlayer decodeAudioSource(Map json) {
|
||||
AudioSourcePlayer decodeAudioSource(AudioSourceMessage audioSourceMessage) {
|
||||
try {
|
||||
switch (json['type']) {
|
||||
case 'progressive':
|
||||
return ProgressiveAudioSourcePlayer(
|
||||
this, json['id'], Uri.parse(json['uri']), json['headers']);
|
||||
case "dash":
|
||||
return DashAudioSourcePlayer(
|
||||
this, json['id'], Uri.parse(json['uri']), json['headers']);
|
||||
case "hls":
|
||||
return HlsAudioSourcePlayer(
|
||||
this, json['id'], Uri.parse(json['uri']), json['headers']);
|
||||
case "concatenating":
|
||||
return ConcatenatingAudioSourcePlayer(
|
||||
this,
|
||||
json['id'],
|
||||
getAudioSources(json['audioSources']),
|
||||
json['useLazyPreparation']);
|
||||
case "clipping":
|
||||
return ClippingAudioSourcePlayer(
|
||||
this,
|
||||
json['id'],
|
||||
getAudioSource(json['audioSource']),
|
||||
Duration(milliseconds: json['start']),
|
||||
Duration(milliseconds: json['end']));
|
||||
case "looping":
|
||||
return LoopingAudioSourcePlayer(this, json['id'],
|
||||
getAudioSource(json['audioSource']), json['count']);
|
||||
default:
|
||||
throw Exception("Unknown AudioSource type: " + json['type']);
|
||||
if (audioSourceMessage is ProgressiveAudioSourceMessage) {
|
||||
return ProgressiveAudioSourcePlayer(this, audioSourceMessage.id,
|
||||
Uri.parse(audioSourceMessage.uri), audioSourceMessage.headers);
|
||||
} else if (audioSourceMessage is DashAudioSourceMessage) {
|
||||
return DashAudioSourcePlayer(this, audioSourceMessage.id,
|
||||
Uri.parse(audioSourceMessage.uri), audioSourceMessage.headers);
|
||||
} else if (audioSourceMessage is HlsAudioSourceMessage) {
|
||||
return HlsAudioSourcePlayer(this, audioSourceMessage.id,
|
||||
Uri.parse(audioSourceMessage.uri), audioSourceMessage.headers);
|
||||
} else if (audioSourceMessage is ConcatenatingAudioSourceMessage) {
|
||||
return ConcatenatingAudioSourcePlayer(
|
||||
this,
|
||||
audioSourceMessage.id,
|
||||
getAudioSources(audioSourceMessage.children),
|
||||
audioSourceMessage.useLazyPreparation);
|
||||
} else if (audioSourceMessage is ClippingAudioSourceMessage) {
|
||||
return ClippingAudioSourcePlayer(
|
||||
this,
|
||||
audioSourceMessage.id,
|
||||
getAudioSource(audioSourceMessage.child),
|
||||
audioSourceMessage.start,
|
||||
audioSourceMessage.end);
|
||||
} else if (audioSourceMessage is LoopingAudioSourceMessage) {
|
||||
return LoopingAudioSourcePlayer(this, audioSourceMessage.id,
|
||||
getAudioSource(audioSourceMessage.child), audioSourceMessage.count);
|
||||
} else {
|
||||
throw Exception("Unknown AudioSource type: $audioSourceMessage");
|
||||
}
|
||||
} catch (e, stacktrace) {
|
||||
print("$stacktrace");
|
||||
|
@ -731,29 +621,6 @@ class ConcatenatingAudioSourcePlayer extends AudioSourcePlayer {
|
|||
return treeIndex;
|
||||
}
|
||||
|
||||
add(AudioSourcePlayer player) {
|
||||
audioSourcePlayers.add(player);
|
||||
_shuffleOrder.add(audioSourcePlayers.length - 1);
|
||||
}
|
||||
|
||||
insert(int index, AudioSourcePlayer player) {
|
||||
audioSourcePlayers.insert(index, player);
|
||||
for (var i = 0; i < audioSourcePlayers.length; i++) {
|
||||
if (_shuffleOrder[i] >= index) {
|
||||
_shuffleOrder[i]++;
|
||||
}
|
||||
}
|
||||
_shuffleOrder.add(index);
|
||||
}
|
||||
|
||||
addAll(List<AudioSourcePlayer> players) {
|
||||
audioSourcePlayers.addAll(players);
|
||||
_shuffleOrder.addAll(
|
||||
List.generate(players.length, (i) => audioSourcePlayers.length + i)
|
||||
.toList()
|
||||
..shuffle());
|
||||
}
|
||||
|
||||
insertAll(int index, List<AudioSourcePlayer> players) {
|
||||
audioSourcePlayers.insertAll(index, players);
|
||||
for (var i = 0; i < audioSourcePlayers.length; i++) {
|
||||
|
@ -765,18 +632,6 @@ class ConcatenatingAudioSourcePlayer extends AudioSourcePlayer {
|
|||
List.generate(players.length, (i) => index + i).toList()..shuffle());
|
||||
}
|
||||
|
||||
removeAt(int index) {
|
||||
audioSourcePlayers.removeAt(index);
|
||||
// 0 1 2 3
|
||||
// 3 2 0 1
|
||||
for (var i = 0; i < audioSourcePlayers.length; i++) {
|
||||
if (_shuffleOrder[i] > index) {
|
||||
_shuffleOrder[i]--;
|
||||
}
|
||||
}
|
||||
_shuffleOrder.removeWhere((i) => i == index);
|
||||
}
|
||||
|
||||
removeRange(int start, int end) {
|
||||
audioSourcePlayers.removeRange(start, end);
|
||||
for (var i = 0; i < audioSourcePlayers.length; i++) {
|
||||
|
@ -791,11 +646,6 @@ class ConcatenatingAudioSourcePlayer extends AudioSourcePlayer {
|
|||
audioSourcePlayers.insert(
|
||||
newIndex, audioSourcePlayers.removeAt(currentIndex));
|
||||
}
|
||||
|
||||
clear() {
|
||||
audioSourcePlayers.clear();
|
||||
_shuffleOrder.clear();
|
||||
}
|
||||
}
|
||||
|
||||
class ClippingAudioSourcePlayer extends IndexedAudioSourcePlayer {
|
||||
|
|
70
just_audio_web/pubspec.lock
Normal file
70
just_audio_web/pubspec.lock
Normal file
|
@ -0,0 +1,70 @@
|
|||
# Generated by pub
|
||||
# See https://dart.dev/tools/pub/glossary#lockfile
|
||||
packages:
|
||||
characters:
|
||||
dependency: transitive
|
||||
description:
|
||||
name: characters
|
||||
url: "https://pub.dartlang.org"
|
||||
source: hosted
|
||||
version: "1.0.0"
|
||||
collection:
|
||||
dependency: transitive
|
||||
description:
|
||||
name: collection
|
||||
url: "https://pub.dartlang.org"
|
||||
source: hosted
|
||||
version: "1.14.13"
|
||||
flutter:
|
||||
dependency: "direct main"
|
||||
description: flutter
|
||||
source: sdk
|
||||
version: "0.0.0"
|
||||
flutter_web_plugins:
|
||||
dependency: "direct main"
|
||||
description: flutter
|
||||
source: sdk
|
||||
version: "0.0.0"
|
||||
just_audio_platform_interface:
|
||||
dependency: "direct main"
|
||||
description:
|
||||
path: "../just_audio_platform_interface"
|
||||
relative: true
|
||||
source: path
|
||||
version: "0.0.1"
|
||||
meta:
|
||||
dependency: "direct main"
|
||||
description:
|
||||
name: meta
|
||||
url: "https://pub.dartlang.org"
|
||||
source: hosted
|
||||
version: "1.1.8"
|
||||
plugin_platform_interface:
|
||||
dependency: transitive
|
||||
description:
|
||||
name: plugin_platform_interface
|
||||
url: "https://pub.dartlang.org"
|
||||
source: hosted
|
||||
version: "1.0.2"
|
||||
sky_engine:
|
||||
dependency: transitive
|
||||
description: flutter
|
||||
source: sdk
|
||||
version: "0.0.99"
|
||||
typed_data:
|
||||
dependency: transitive
|
||||
description:
|
||||
name: typed_data
|
||||
url: "https://pub.dartlang.org"
|
||||
source: hosted
|
||||
version: "1.2.0"
|
||||
vector_math:
|
||||
dependency: transitive
|
||||
description:
|
||||
name: vector_math
|
||||
url: "https://pub.dartlang.org"
|
||||
source: hosted
|
||||
version: "2.0.8"
|
||||
sdks:
|
||||
dart: ">=2.9.0-14.0.dev <3.0.0"
|
||||
flutter: ">=1.12.13+hotfix.5"
|
25
just_audio_web/pubspec.yaml
Normal file
25
just_audio_web/pubspec.yaml
Normal file
|
@ -0,0 +1,25 @@
|
|||
name: just_audio_web
|
||||
description: Web platform implementation of just_audio
|
||||
homepage: https://github.com/ryanheise/just_audio/tree/master/just_audio_web
|
||||
version: 0.0.1
|
||||
|
||||
flutter:
|
||||
plugin:
|
||||
platforms:
|
||||
web:
|
||||
pluginClass: JustAudioPlugin
|
||||
fileName: just_audio_web.dart
|
||||
|
||||
dependencies:
|
||||
#just_audio_platform_interface: ^0.0.1
|
||||
just_audio_platform_interface:
|
||||
path: ../just_audio_platform_interface
|
||||
flutter:
|
||||
sdk: flutter
|
||||
flutter_web_plugins:
|
||||
sdk: flutter
|
||||
meta: ^1.1.8
|
||||
|
||||
environment:
|
||||
sdk: ">=2.7.0 <3.0.0"
|
||||
flutter: ">=1.12.13+hotfix.5"
|
Loading…
Add table
Reference in a new issue