use std::collections::HashMap; use std::io::Cursor; use std::ops::{Deref, DerefMut}; use std::sync::Arc; pub use alto::efx; pub use alto::Context; pub use alto::Device; pub use alto::Source; use alto::{efx::AuxEffectSlot, ContextAttrs, SourceState}; use alto::{Alto, Mono, StaticSource, Stereo}; use bevy::{ asset::{AssetLoader, HandleId, LoadContext, LoadedAsset}, prelude::*, reflect::TypeUuid, utils::BoxedFuture, }; use lewton::inside_ogg::OggStreamReader; #[derive(Clone, Debug, TypeUuid)] #[uuid = "aa22d11e-3bed-11eb-8708-00155dea3db9"] pub struct Buffer { samples: Vec, sample_rate: i32, channels: u16, } #[derive(Clone, Copy, Debug, Default)] pub struct BufferAssetLoader; impl AssetLoader for BufferAssetLoader { fn load<'a>( &'a self, bytes: &'a [u8], load_context: &'a mut LoadContext, ) -> BoxedFuture<'a, Result<(), anyhow::Error>> { Box::pin(async move { let cursor = Cursor::new(bytes.to_vec()); let buffer: Option = match load_context.path().extension().unwrap().to_str().unwrap() { "flac" => { let reader = claxon::FlacReader::new(cursor); if let Ok(mut reader) = reader { let mut samples: Vec = vec![]; for sample in reader.samples() { if let Ok(sample) = sample { samples.push(sample as i16); } } let info = reader.streaminfo(); Some(Buffer { samples, sample_rate: info.sample_rate as i32, channels: info.channels as u16, }) } else { None } } "ogg" => { let mut stream = OggStreamReader::new(cursor)?; let mut samples: Vec = vec![]; while let Some(pck_samples) = &mut stream.read_dec_packet_itl()? { samples.append(pck_samples); } Some(Buffer { samples, channels: stream.ident_hdr.audio_channels as u16, sample_rate: stream.ident_hdr.audio_sample_rate as i32, }) } "wav" => { let reader = hound::WavReader::new(cursor); if let Ok(mut reader) = reader { let mut samples: Vec = vec![]; for sample in reader.samples::() { if let Ok(sample) = sample { samples.push(sample); } } Some(Buffer { samples, sample_rate: reader.spec().sample_rate as i32, channels: reader.spec().channels, }) } else { None } } _ => None, }; if let Some(buffer) = buffer { load_context.set_default_asset(LoadedAsset::new(buffer)); } Ok(()) }) } fn extensions(&self) -> &[&str] { &["flac", "ogg", "wav"] } } // TODO: Make non-public when we have multi-stage asset loading. #[derive(Default)] pub struct Buffers(pub HashMap>); fn buffer_creation( context: Res, mut buffers: ResMut, mut events: EventReader>, assets: Res>, ) { for event in events.iter() { match event { AssetEvent::Created { handle } => { if let Some(buffer) = assets.get(handle) { let buffer = match buffer.channels { 1 => { context.new_buffer::, _>(&buffer.samples, buffer.sample_rate) } 2 => context .new_buffer::, _>(&buffer.samples, buffer.sample_rate), _ => { panic!("Unsupported channel count"); } }; if let Ok(buffer) = buffer { buffers.0.insert(handle.id, Arc::new(buffer)); } } } AssetEvent::Modified { handle: _ } => {} AssetEvent::Removed { handle } => { buffers.0.remove(&handle.id); } } } } #[derive(Clone, Copy, Debug, PartialEq, Reflect)] pub enum SoundState { Stopped, Playing, Paused, } impl Default for SoundState { fn default() -> Self { SoundState::Stopped } } #[derive(Reflect)] pub struct Sound { pub buffer: Handle, pub state: SoundState, pub gain: f32, pub looping: bool, pub pitch: f32, pub reference_distance: f32, pub max_distance: f32, pub rolloff_factor: f32, pub bypass_global_effects: bool, #[reflect(ignore)] pub source: Option, } impl Default for Sound { fn default() -> Self { Self { buffer: Default::default(), state: Default::default(), gain: 1., looping: false, pitch: 1., reference_distance: 1., max_distance: f32::MAX, rolloff_factor: 1., bypass_global_effects: false, source: None, } } } #[derive(Default)] pub struct GlobalEffects(Vec); impl Deref for GlobalEffects { type Target = Vec; fn deref(&self) -> &Self::Target { &self.0 } } impl DerefMut for GlobalEffects { fn deref_mut(&mut self) -> &mut Self::Target { &mut self.0 } } fn update_source_position( source: &mut StaticSource, transform: Option<&Transform>, global_transform: Option<&GlobalTransform>, ) { let translation = global_transform .map(|v| v.translation) .or_else(|| transform.map(|v| v.translation)); if let Some(translation) = translation { source.set_relative(false); source .set_position([translation.x, translation.y, translation.z]) .unwrap(); } else { //println!("No translation: {:?}, {:?}", transform, global_transform); source.set_relative(true); source.set_position([0., 0., 0.]).unwrap(); } } fn source_update( context: Res, buffers: Res, mut global_effects: ResMut, mut query: Query<(&mut Sound, Option<&Transform>, Option<&GlobalTransform>)>, ) { for (mut sound, transform, global_transform) in query.iter_mut() { let state = sound.state; match state { SoundState::Stopped => { if let Some(source) = sound.source.as_mut() { source.stop(); sound.source = None; } } SoundState::Playing => { if let Some(source) = sound.source.as_mut() { let source_state = source.state(); if source_state == SourceState::Paused { source.play(); } } else { let mut source = context.new_static_source().unwrap(); if let Some(buffer) = buffers.0.get(&sound.buffer.id) { source.set_buffer(buffer.clone()).unwrap(); } update_source_position(&mut source, transform, global_transform); source.play(); sound.source = Some(source); } } SoundState::Paused => { if let Some(source) = sound.source.as_mut() { source.pause(); } else { let mut source = context.new_static_source().unwrap(); if let Some(buffer) = buffers.0.get(&sound.buffer.id) { source.set_buffer(buffer.clone()).unwrap(); } source.pause(); sound.source = Some(source); } } } let state = sound.state; let gain = sound.gain; let looping = sound.looping; let pitch = sound.pitch; let reference_distance = sound.reference_distance; let max_distance = sound.max_distance; let rolloff_factor = sound.rolloff_factor; let bypass_global_effects = sound.bypass_global_effects; let source_state = if let Some(source) = &sound.source { Some(source.state()) } else { None }; if let Some(source_state) = source_state { sound.state = match source_state { SourceState::Initial => SoundState::Stopped, SourceState::Playing => SoundState::Playing, SourceState::Paused => SoundState::Paused, SourceState::Stopped => SoundState::Stopped, SourceState::Unknown(_) => SoundState::Stopped, }; } else { sound.state = SoundState::Stopped; } if let Some(source) = sound.source.as_mut() { if state != SoundState::Stopped { source.set_gain(gain).unwrap(); source.set_looping(looping); source.set_pitch(pitch).unwrap(); source.set_reference_distance(reference_distance).unwrap(); source.set_max_distance(max_distance).unwrap(); source.set_rolloff_factor(rolloff_factor).unwrap(); update_source_position(source, transform, global_transform); if !bypass_global_effects { for (send, effect) in global_effects.iter_mut().enumerate() { source.set_aux_send(send as i32, effect).unwrap(); } } } else { sound.source = None; } } } } impl Sound { pub fn stop(&mut self) { if let Some(source) = self.source.as_mut() { source.stop(); } self.state = SoundState::Stopped; self.source = None; } pub fn play(&mut self) { if let Some(source) = self.source.as_mut() { source.play(); } self.state = SoundState::Playing; } pub fn pause(&mut self) { if let Some(source) = self.source.as_mut() { source.pause(); } self.state = SoundState::Paused; } } #[derive(Clone, Copy, Debug, Default, Reflect)] #[reflect(Component)] pub struct Listener; fn listener_update( context: ResMut, query: Query<(&Listener, Option<&Transform>, Option<&GlobalTransform>)>, ) { for (_, transform, global_transform) in query.iter() { let transform: Option = global_transform .map(|v| { let transform: Transform = (*v).into(); transform }) .or_else(|| transform.cloned()); if let Some(transform) = transform { let look = transform.local_x(); let up = transform.local_z(); context .set_position([ transform.translation.x, transform.translation.y, transform.translation.z, ]) .unwrap(); context .set_orientation(([look.x, look.y, look.z], [up.x, up.y, up.z])) .unwrap(); } else { context.set_position([0., 0., 0.]).unwrap(); context .set_orientation(([0., 0., 1.], [0., 1., 0.])) .unwrap(); } } } #[derive(Clone, Copy, Debug, Default)] pub struct OpenAlConfig { pub soft_hrtf: bool, } pub struct OpenAlPlugin; impl Plugin for OpenAlPlugin { fn build(&self, app: &mut AppBuilder) { if !app.world().contains_resource::() { app.insert_resource(OpenAlConfig::default()); } let config = app.world().get_resource::().unwrap().clone(); let al = Alto::load_default().expect("Could not load alto"); let device = al.open(None).expect("Could not open device"); let mut context_attrs = ContextAttrs::default(); if config.soft_hrtf { context_attrs.soft_hrtf = Some(true); } let context = device .new_context(Some(context_attrs)) .expect("Could not create context"); app.add_asset::() .init_asset_loader::() .insert_non_send_resource(device) .insert_resource(context) .insert_resource(Buffers::default()) .insert_resource(GlobalEffects::default()) .register_type::() .add_system(buffer_creation.system()) .add_system(source_update.system()) .add_system(listener_update.system()); } }