}
int clipOut = 0;
+ int duration = 0;
if (xml.hasAttribute("out")) clipOut = xml.attribute("out").toInt();
// setup length here as otherwise default length (currently 15000 frames in MLT) will be taken even if outpoint is larger
if (type == COLOR || type == TEXT || type == IMAGE || type == SLIDESHOW) {
int length;
if (xml.hasAttribute("length")) {
+ if (clipOut > 0) duration = clipOut + 1;
length = xml.attribute("length").toInt();
clipOut = length - 1;
}
int frameNumber = xml.attribute("thumbnail", "0").toInt();
if (frameNumber != 0) producer->seek(frameNumber);
- filePropertyMap["duration"] = QString::number(producer->get_playtime());
- //kDebug() << "/////// PRODUCER: " << url.path() << " IS: " << producer.get_playtime();
+ filePropertyMap["duration"] = QString::number(duration > 0 ? duration : producer->get_playtime());
+ //kDebug() << "/////// PRODUCER: " << url.path() << " IS: " << producer->get_playtime();
Mlt::Frame *frame = producer->get_frame();
Mlt::Tractor tractor(service);
Mlt::Producer trackProducer(tractor.track(track));
+ // Make sure muting will not produce problems with our audio mixing transition,
+ // because audio mixing is done between each track and the lowest one
+ bool audioMixingBroken = false;
+ if (mute && trackProducer.get_int("hide") < 2 ) {
+ // We mute a track with sound
+ if (track == getLowestNonMutedAudioTrack(tractor)) audioMixingBroken = true;
+ }
+ else if (!mute && trackProducer.get_int("hide") > 1 ) {
+ // We un-mute a previously muted track
+ if (track < getLowestNonMutedAudioTrack(tractor)) audioMixingBroken = true;
+ }
+
if (mute) {
if (blind) trackProducer.set("hide", 3);
else trackProducer.set("hide", 2);
} else {
trackProducer.set("hide", 0);
}
+ if (audioMixingBroken) fixAudioMixing(tractor);
+
tractor.multitrack()->refresh();
tractor.refresh();
refresh();
}
+int Render::getLowestNonMutedAudioTrack(Mlt::Tractor tractor)
+{
+ for (int i = 1; i < tractor.count(); i++) {
+ Mlt::Producer trackProducer(tractor.track(i));
+ if (trackProducer.get_int("hide") < 2) return i;
+ }
+ return tractor.count() - 1;
+}
+
+void Render::fixAudioMixing(Mlt::Tractor tractor)
+{
+ // Make sure the audio mixing transitions are applied to the lowest audible (non muted) track
+ int lowestTrack = getLowestNonMutedAudioTrack(tractor);
+
+ mlt_service serv = m_mltProducer->parent().get_service();
+ mlt_service_lock(serv);
+ m_isBlocked++;
+
+ mlt_service nextservice = mlt_service_get_producer(serv);
+ mlt_properties properties = MLT_SERVICE_PROPERTIES(nextservice);
+ QString mlt_type = mlt_properties_get(properties, "mlt_type");
+ QString resource = mlt_properties_get(properties, "mlt_service");
+
+ while (mlt_type == "transition") {
+ mlt_transition tr = (mlt_transition) nextservice;
+ if (resource == "mix") {
+ mlt_properties transproperties = MLT_TRANSITION_PROPERTIES(tr);
+ mlt_properties_set_int(transproperties, "a_track", lowestTrack);
+ }
+ nextservice = mlt_service_producer(nextservice);
+ if (nextservice == NULL) break;
+ properties = MLT_SERVICE_PROPERTIES(nextservice);
+ mlt_type = mlt_properties_get(properties, "mlt_type");
+ resource = mlt_properties_get(properties, "mlt_service");
+ }
+ mlt_service_unlock(serv);
+ m_isBlocked--;
+}
bool Render::mltResizeClipCrop(ItemInfo info, GenTime diff)
{