compile_obj_transform( obj, audio.transform )
audio.clip_start = audio_clip_count
audio.clip_count = len(obj_data.files)
+ audio_clip_count += audio.clip_count
audio.max_channels = obj_data.max_channels
audio.volume = obj_data.volume
elif obj_data.formato == '1': audio.flags |= 0x400
elif obj_data.formato == '2': audio.flags |= 0x1000
+ audio.channel_behaviour = int(obj_data.channel_behaviour)
+ if audio.channel_behaviour >= 1:#{
+ audio.group = obj_data.group
+ #}
+ if audio.channel_behaviour == 2:#{
+ audio.crossfade = obj_data.transition_duration
+ #}
+ audio.probability_curve = int(obj_data.probability_curve)
+
for ci in range(audio.clip_count):#{
entry = obj_data.files[ci]
clip = ent_audio_clip()
flag_loop: bpy.props.BoolProperty( name="Loop",default=False )
flag_auto: bpy.props.BoolProperty( name="Play at start",default=False )
flag_nodoppler: bpy.props.BoolProperty( name="No Doppler",default=False )
+
+ group: bpy.props.IntProperty( name="Group ID", default=0 )
formato: bpy.props.EnumProperty(
name="Format",
items=[('0','Uncompressed Mono',''),
c.prop( data[0], 'max_channels' )
c = split.column()
c.prop( data[0], 'channel_behaviour', text='Behaviour' )
+ if data[0].channel_behaviour >= '1':
+ box.prop( data[0], 'group' )
if data[0].channel_behaviour == '2':
box.prop( data[0], 'transition_duration' )
cv_ent_volume( obj )
#}
elif ent_type == 'ent_audio':#{
- cv_draw_sphere( obj.location, obj.scale[0], (1,1,0) )
+ if obj.SR_data.ent_audio[0].flag_3d:
+ cv_draw_sphere( obj.location, obj.scale[0], (1,1,0) )
#}
#}
#}