diff options
author | Matt A. Tobin <mattatobin@localhost.localdomain> | 2018-02-02 04:16:08 -0500 |
---|---|---|
committer | Matt A. Tobin <mattatobin@localhost.localdomain> | 2018-02-02 04:16:08 -0500 |
commit | 5f8de423f190bbb79a62f804151bc24824fa32d8 (patch) | |
tree | 10027f336435511475e392454359edea8e25895d /dom/media/webrtc/MediaEngineCameraVideoSource.cpp | |
parent | 49ee0794b5d912db1f95dce6eb52d781dc210db5 (diff) | |
download | UXP-5f8de423f190bbb79a62f804151bc24824fa32d8.tar UXP-5f8de423f190bbb79a62f804151bc24824fa32d8.tar.gz UXP-5f8de423f190bbb79a62f804151bc24824fa32d8.tar.lz UXP-5f8de423f190bbb79a62f804151bc24824fa32d8.tar.xz UXP-5f8de423f190bbb79a62f804151bc24824fa32d8.zip |
Add m-esr52 at 52.6.0
Diffstat (limited to 'dom/media/webrtc/MediaEngineCameraVideoSource.cpp')
-rw-r--r-- | dom/media/webrtc/MediaEngineCameraVideoSource.cpp | 418 |
1 files changed, 418 insertions, 0 deletions
diff --git a/dom/media/webrtc/MediaEngineCameraVideoSource.cpp b/dom/media/webrtc/MediaEngineCameraVideoSource.cpp new file mode 100644 index 000000000..a0f31d937 --- /dev/null +++ b/dom/media/webrtc/MediaEngineCameraVideoSource.cpp @@ -0,0 +1,418 @@ +/* This Source Code Form is subject to the terms of the Mozilla Public + * License, v. 2.0. If a copy of the MPL was not distributed with this file, + * You can obtain one at http://mozilla.org/MPL/2.0/. */ + +#include "MediaEngineCameraVideoSource.h" + +#include <limits> + +namespace mozilla { + +using namespace mozilla::gfx; +using namespace mozilla::dom; + +extern LogModule* GetMediaManagerLog(); +#define LOG(msg) MOZ_LOG(GetMediaManagerLog(), mozilla::LogLevel::Debug, msg) +#define LOGFRAME(msg) MOZ_LOG(GetMediaManagerLog(), mozilla::LogLevel::Verbose, msg) + +// guts for appending data to the MSG track +bool MediaEngineCameraVideoSource::AppendToTrack(SourceMediaStream* aSource, + layers::Image* aImage, + TrackID aID, + StreamTime delta, + const PrincipalHandle& aPrincipalHandle) +{ + MOZ_ASSERT(aSource); + + VideoSegment segment; + RefPtr<layers::Image> image = aImage; + IntSize size(image ? mWidth : 0, image ? mHeight : 0); + segment.AppendFrame(image.forget(), delta, size, aPrincipalHandle); + + // This is safe from any thread, and is safe if the track is Finished + // or Destroyed. + // This can fail if either a) we haven't added the track yet, or b) + // we've removed or finished the track. + return aSource->AppendToTrack(aID, &(segment)); +} + +// Sub-classes (B2G or desktop) should overload one of both of these two methods +// to provide capabilities +size_t +MediaEngineCameraVideoSource::NumCapabilities() const +{ + return mHardcodedCapabilities.Length(); +} + +void +MediaEngineCameraVideoSource::GetCapability(size_t aIndex, + webrtc::CaptureCapability& aOut) const +{ + MOZ_ASSERT(aIndex < mHardcodedCapabilities.Length()); + aOut = mHardcodedCapabilities.SafeElementAt(aIndex, webrtc::CaptureCapability()); +} + +uint32_t +MediaEngineCameraVideoSource::GetFitnessDistance( + const webrtc::CaptureCapability& aCandidate, + const NormalizedConstraintSet &aConstraints, + const nsString& aDeviceId) const +{ + // Treat width|height|frameRate == 0 on capability as "can do any". + // This allows for orthogonal capabilities that are not in discrete steps. + + uint64_t distance = + uint64_t(FitnessDistance(aDeviceId, aConstraints.mDeviceId)) + + uint64_t(FitnessDistance(mFacingMode, aConstraints.mFacingMode)) + + uint64_t(aCandidate.width? FitnessDistance(int32_t(aCandidate.width), + aConstraints.mWidth) : 0) + + uint64_t(aCandidate.height? FitnessDistance(int32_t(aCandidate.height), + aConstraints.mHeight) : 0) + + uint64_t(aCandidate.maxFPS? FitnessDistance(double(aCandidate.maxFPS), + aConstraints.mFrameRate) : 0); + return uint32_t(std::min(distance, uint64_t(UINT32_MAX))); +} + +// Find best capability by removing inferiors. May leave >1 of equal distance + +/* static */ void +MediaEngineCameraVideoSource::TrimLessFitCandidates(CapabilitySet& set) { + uint32_t best = UINT32_MAX; + for (auto& candidate : set) { + if (best > candidate.mDistance) { + best = candidate.mDistance; + } + } + for (size_t i = 0; i < set.Length();) { + if (set[i].mDistance > best) { + set.RemoveElementAt(i); + } else { + ++i; + } + } + MOZ_ASSERT(set.Length()); +} + +// GetBestFitnessDistance returns the best distance the capture device can offer +// as a whole, given an accumulated number of ConstraintSets. +// Ideal values are considered in the first ConstraintSet only. +// Plain values are treated as Ideal in the first ConstraintSet. +// Plain values are treated as Exact in subsequent ConstraintSets. +// Infinity = UINT32_MAX e.g. device cannot satisfy accumulated ConstraintSets. +// A finite result may be used to calculate this device's ranking as a choice. + +uint32_t +MediaEngineCameraVideoSource::GetBestFitnessDistance( + const nsTArray<const NormalizedConstraintSet*>& aConstraintSets, + const nsString& aDeviceId) const +{ + size_t num = NumCapabilities(); + + CapabilitySet candidateSet; + for (size_t i = 0; i < num; i++) { + candidateSet.AppendElement(i); + } + + bool first = true; + for (const NormalizedConstraintSet* ns : aConstraintSets) { + for (size_t i = 0; i < candidateSet.Length(); ) { + auto& candidate = candidateSet[i]; + webrtc::CaptureCapability cap; + GetCapability(candidate.mIndex, cap); + uint32_t distance = GetFitnessDistance(cap, *ns, aDeviceId); + if (distance == UINT32_MAX) { + candidateSet.RemoveElementAt(i); + } else { + ++i; + if (first) { + candidate.mDistance = distance; + } + } + } + first = false; + } + if (!candidateSet.Length()) { + return UINT32_MAX; + } + TrimLessFitCandidates(candidateSet); + return candidateSet[0].mDistance; +} + +void +MediaEngineCameraVideoSource::LogConstraints( + const NormalizedConstraintSet& aConstraints) +{ + auto& c = aConstraints; + LOG(((c.mWidth.mIdeal.isSome()? + "Constraints: width: { min: %d, max: %d, ideal: %d }" : + "Constraints: width: { min: %d, max: %d }"), + c.mWidth.mMin, c.mWidth.mMax, + c.mWidth.mIdeal.valueOr(0))); + LOG(((c.mHeight.mIdeal.isSome()? + " height: { min: %d, max: %d, ideal: %d }" : + " height: { min: %d, max: %d }"), + c.mHeight.mMin, c.mHeight.mMax, + c.mHeight.mIdeal.valueOr(0))); + LOG(((c.mFrameRate.mIdeal.isSome()? + " frameRate: { min: %f, max: %f, ideal: %f }" : + " frameRate: { min: %f, max: %f }"), + c.mFrameRate.mMin, c.mFrameRate.mMax, + c.mFrameRate.mIdeal.valueOr(0))); +} + +void +MediaEngineCameraVideoSource::LogCapability(const char* aHeader, + const webrtc::CaptureCapability &aCapability, uint32_t aDistance) +{ + // RawVideoType and VideoCodecType media/webrtc/trunk/webrtc/common_types.h + static const char* const types[] = { + "I420", + "YV12", + "YUY2", + "UYVY", + "IYUV", + "ARGB", + "RGB24", + "RGB565", + "ARGB4444", + "ARGB1555", + "MJPEG", + "NV12", + "NV21", + "BGRA", + "Unknown type" + }; + + static const char* const codec[] = { + "VP8", + "VP9", + "H264", + "I420", + "RED", + "ULPFEC", + "Generic codec", + "Unknown codec" + }; + + LOG(("%s: %4u x %4u x %2u maxFps, %s, %s. Distance = %lu", + aHeader, aCapability.width, aCapability.height, aCapability.maxFPS, + types[std::min(std::max(uint32_t(0), uint32_t(aCapability.rawType)), + uint32_t(sizeof(types) / sizeof(*types) - 1))], + codec[std::min(std::max(uint32_t(0), uint32_t(aCapability.codecType)), + uint32_t(sizeof(codec) / sizeof(*codec) - 1))], + aDistance)); +} + +bool +MediaEngineCameraVideoSource::ChooseCapability( + const NormalizedConstraints &aConstraints, + const MediaEnginePrefs &aPrefs, + const nsString& aDeviceId) +{ + if (MOZ_LOG_TEST(GetMediaManagerLog(), LogLevel::Debug)) { + LOG(("ChooseCapability: prefs: %dx%d @%d-%dfps", + aPrefs.GetWidth(), aPrefs.GetHeight(), + aPrefs.mFPS, aPrefs.mMinFPS)); + LogConstraints(aConstraints); + if (aConstraints.mAdvanced.size()) { + LOG(("Advanced array[%u]:", aConstraints.mAdvanced.size())); + for (auto& advanced : aConstraints.mAdvanced) { + LogConstraints(advanced); + } + } + } + + size_t num = NumCapabilities(); + + CapabilitySet candidateSet; + for (size_t i = 0; i < num; i++) { + candidateSet.AppendElement(i); + } + + // First, filter capabilities by required constraints (min, max, exact). + + for (size_t i = 0; i < candidateSet.Length();) { + auto& candidate = candidateSet[i]; + webrtc::CaptureCapability cap; + GetCapability(candidate.mIndex, cap); + candidate.mDistance = GetFitnessDistance(cap, aConstraints, aDeviceId); + LogCapability("Capability", cap, candidate.mDistance); + if (candidate.mDistance == UINT32_MAX) { + candidateSet.RemoveElementAt(i); + } else { + ++i; + } + } + + if (!candidateSet.Length()) { + LOG(("failed to find capability match from %d choices",num)); + return false; + } + + // Filter further with all advanced constraints (that don't overconstrain). + + for (const auto &cs : aConstraints.mAdvanced) { + CapabilitySet rejects; + for (size_t i = 0; i < candidateSet.Length();) { + auto& candidate = candidateSet[i]; + webrtc::CaptureCapability cap; + GetCapability(candidate.mIndex, cap); + if (GetFitnessDistance(cap, cs, aDeviceId) == UINT32_MAX) { + rejects.AppendElement(candidate); + candidateSet.RemoveElementAt(i); + } else { + ++i; + } + } + if (!candidateSet.Length()) { + candidateSet.AppendElements(Move(rejects)); + } + } + MOZ_ASSERT(candidateSet.Length(), + "advanced constraints filtering step can't reduce candidates to zero"); + + // Remaining algorithm is up to the UA. + + TrimLessFitCandidates(candidateSet); + + // Any remaining multiples all have the same distance. A common case of this + // occurs when no ideal is specified. Lean toward defaults. + uint32_t sameDistance = candidateSet[0].mDistance; + { + MediaTrackConstraintSet prefs; + prefs.mWidth.SetAsLong() = aPrefs.GetWidth(); + prefs.mHeight.SetAsLong() = aPrefs.GetHeight(); + prefs.mFrameRate.SetAsDouble() = aPrefs.mFPS; + NormalizedConstraintSet normPrefs(prefs, false); + + for (auto& candidate : candidateSet) { + webrtc::CaptureCapability cap; + GetCapability(candidate.mIndex, cap); + candidate.mDistance = GetFitnessDistance(cap, normPrefs, aDeviceId); + } + TrimLessFitCandidates(candidateSet); + } + + // Any remaining multiples all have the same distance, but may vary on + // format. Some formats are more desirable for certain use like WebRTC. + // E.g. I420 over RGB24 can remove a needless format conversion. + + bool found = false; + for (auto& candidate : candidateSet) { + webrtc::CaptureCapability cap; + GetCapability(candidate.mIndex, cap); + if (cap.rawType == webrtc::RawVideoType::kVideoI420 || + cap.rawType == webrtc::RawVideoType::kVideoYUY2 || + cap.rawType == webrtc::RawVideoType::kVideoYV12) { + mCapability = cap; + found = true; + break; + } + } + if (!found) { + GetCapability(candidateSet[0].mIndex, mCapability); + } + + LogCapability("Chosen capability", mCapability, sameDistance); + return true; +} + +void +MediaEngineCameraVideoSource::SetName(nsString aName) +{ + mDeviceName = aName; + bool hasFacingMode = false; + VideoFacingModeEnum facingMode = VideoFacingModeEnum::User; + + // Set facing mode based on device name. +#if defined(ANDROID) && !defined(MOZ_WIDGET_GONK) + // Names are generated. Example: "Camera 0, Facing back, Orientation 90" + // + // See media/webrtc/trunk/webrtc/modules/video_capture/android/java/src/org/ + // webrtc/videoengine/VideoCaptureDeviceInfoAndroid.java + + if (aName.Find(NS_LITERAL_STRING("Facing back")) != kNotFound) { + hasFacingMode = true; + facingMode = VideoFacingModeEnum::Environment; + } else if (aName.Find(NS_LITERAL_STRING("Facing front")) != kNotFound) { + hasFacingMode = true; + facingMode = VideoFacingModeEnum::User; + } +#endif // ANDROID +#ifdef XP_MACOSX + // Kludge to test user-facing cameras on OSX. + if (aName.Find(NS_LITERAL_STRING("Face")) != -1) { + hasFacingMode = true; + facingMode = VideoFacingModeEnum::User; + } +#endif +#ifdef XP_WIN + // The cameras' name of Surface book are "Microsoft Camera Front" and + // "Microsoft Camera Rear" respectively. + + if (aName.Find(NS_LITERAL_STRING("Front")) != kNotFound) { + hasFacingMode = true; + facingMode = VideoFacingModeEnum::User; + } else if (aName.Find(NS_LITERAL_STRING("Rear")) != kNotFound) { + hasFacingMode = true; + facingMode = VideoFacingModeEnum::Environment; + } +#endif // WINDOWS + if (hasFacingMode) { + mFacingMode.Assign(NS_ConvertUTF8toUTF16( + VideoFacingModeEnumValues::strings[uint32_t(facingMode)].value)); + } else { + mFacingMode.Truncate(); + } +} + +void +MediaEngineCameraVideoSource::GetName(nsAString& aName) const +{ + aName = mDeviceName; +} + +void +MediaEngineCameraVideoSource::SetUUID(const char* aUUID) +{ + mUniqueId.Assign(aUUID); +} + +void +MediaEngineCameraVideoSource::GetUUID(nsACString& aUUID) const +{ + aUUID = mUniqueId; +} + +const nsCString& +MediaEngineCameraVideoSource::GetUUID() const +{ + return mUniqueId; +} + +void +MediaEngineCameraVideoSource::SetDirectListeners(bool aHasDirectListeners) +{ + LOG((__FUNCTION__)); + mHasDirectListeners = aHasDirectListeners; +} + +bool operator == (const webrtc::CaptureCapability& a, + const webrtc::CaptureCapability& b) +{ + return a.width == b.width && + a.height == b.height && + a.maxFPS == b.maxFPS && + a.rawType == b.rawType && + a.codecType == b.codecType && + a.expectedCaptureDelay == b.expectedCaptureDelay && + a.interlaced == b.interlaced; +}; + +bool operator != (const webrtc::CaptureCapability& a, + const webrtc::CaptureCapability& b) +{ + return !(a == b); +} + +} // namespace mozilla |