feat: integrate FFmpeg for audio processing and add WASM support for fingerprint generation

This commit is contained in:
Chigozirim Igweamaka 2025-04-01 17:44:54 +01:00
parent 34122d10a5
commit 66b0071698

View file

@ -9,20 +9,28 @@ import { ToastContainer, toast, Slide } from "react-toastify";
import "react-toastify/dist/ReactToastify.css"; import "react-toastify/dist/ReactToastify.css";
import { MediaRecorder, register } from "extendable-media-recorder"; import { MediaRecorder, register } from "extendable-media-recorder";
import { connect } from "extendable-media-recorder-wav-encoder"; import { connect } from "extendable-media-recorder-wav-encoder";
import { FFmpeg } from '@ffmpeg/ffmpeg';
import { fetchFile } from '@ffmpeg/util';
import AnimatedNumber from "./components/AnimatedNumber"; import AnimatedNumber from "./components/AnimatedNumber";
const server = process.env.REACT_APP_BACKEND_URL || "http://localhost:5000"; const server = process.env.REACT_APP_BACKEND_URL || "http://localhost:5000";
// https://seek-tune-rq4gn.ondigitalocean.app/
var socket = io(server); var socket = io(server);
function App() { function App() {
let ffmpegLoaded = false;
const ffmpeg = new FFmpeg();
const uploadRecording = true
const isPhone = window.innerWidth <= 550
const [stream, setStream] = useState(); const [stream, setStream] = useState();
const [matches, setMatches] = useState([]); const [matches, setMatches] = useState([]);
const [totalSongs, setTotalSongs] = useState(10); const [totalSongs, setTotalSongs] = useState(10);
const [isListening, setisListening] = useState(false); const [isListening, setisListening] = useState(false);
const [audioInput, setAudioInput] = useState("device"); // or "mic" const [audioInput, setAudioInput] = useState("device"); // or "mic"
const [isPhone, setIsPhone] = useState(window.innerWidth <= 550); const [genFingerprint, setGenFingerprint] = useState(null);
const [registeredMediaEncoder, setRegisteredMediaEncoder] = useState(false); const [registeredMediaEncoder, setRegisteredMediaEncoder] = useState(false);
const streamRef = useRef(stream); const streamRef = useRef(stream);
@ -78,8 +86,38 @@ function App() {
return () => clearInterval(intervalId); return () => clearInterval(intervalId);
}, []); }, []);
useEffect(() => {
(async () => {
try {
const go = new window.Go();
const result = await WebAssembly.instantiateStreaming(
fetch("/main.wasm"),
go.importObject
);
go.run(result.instance);
if (typeof window.generateFingerprint === "function") {
setGenFingerprint(() => window.generateFingerprint);
}
} catch (error) {
console.error("Error loading WASM:", error);
}
})();
}, []);
async function record() { async function record() {
try { try {
if (!genFingerprint) {
console.error("WASM is not loaded yet.");
return;
}
if (!ffmpegLoaded) {
await ffmpeg.load();
ffmpegLoaded = true;
}
const mediaDevice = const mediaDevice =
audioInput === "device" audioInput === "device"
? navigator.mediaDevices.getDisplayMedia.bind(navigator.mediaDevices) ? navigator.mediaDevices.getDisplayMedia.bind(navigator.mediaDevices)
@ -113,33 +151,6 @@ function App() {
track.stop(); track.stop();
} }
/** Attempt to change sampleRate
const audioContext = new AudioContext({
sampleRate: 44100,
});
const mediaStreamAudioSourceNode = new MediaStreamAudioSourceNode(
audioContext,
{ mediaStream: audioStream }
);
const mediaStreamAudioDestinationNode =
new MediaStreamAudioDestinationNode(audioContext, {
channelCount: 1,
});
mediaStreamAudioSourceNode.connect(mediaStreamAudioDestinationNode);
const mediaRecorder = new MediaRecorder(
mediaStreamAudioDestinationNode.stream,
{ mimeType: "audio/wav" }
);
const settings = mediaStreamAudioDestinationNode.stream
.getAudioTracks()[0]
.getSettings();
console.log("Settings: ", settings);
*/
const mediaRecorder = new MediaRecorder(audioStream, { const mediaRecorder = new MediaRecorder(audioStream, {
mimeType: "audio/wav", mimeType: "audio/wav",
}); });
@ -158,45 +169,77 @@ function App() {
mediaRecorder.stop(); mediaRecorder.stop();
}, 20000); }, 20000);
mediaRecorder.addEventListener("stop", () => { mediaRecorder.addEventListener("stop", async () => {
const blob = new Blob(chunks, { type: "audio/wav" }); const blob = new Blob(chunks, { type: "audio/wav" });
const reader = new FileReader();
cleanUp(); cleanUp();
// downloadRecording(blob);
reader.readAsArrayBuffer(blob); const inputFile = 'input.wav';
reader.onload = async (event) => { const outputFile = 'output_mono.wav';
const arrayBuffer = event.target.result;
// get record duration // Convert audio to mono with a sample rate of 44100 Hz
const arrayBufferCopy = arrayBuffer.slice(0); await ffmpeg.writeFile(inputFile, await fetchFile(blob))
const audioContext = new AudioContext(); const exitCode = await ffmpeg.exec([
const audioBufferDecoded = await audioContext.decodeAudioData( '-i', inputFile,
arrayBufferCopy '-c', 'pcm_s16le',
); '-ar', '44100',
const recordDuration = audioBufferDecoded.duration; '-ac', '1',
'-f', 'wav',
var binary = ""; outputFile
var bytes = new Uint8Array(arrayBuffer); ]);
var len = bytes.byteLength; if (exitCode !== 0) {
for (var i = 0; i < len; i++) { throw new Error(`FFmpeg exec failed with exit code: ${exitCode}`);
binary += String.fromCharCode(bytes[i]);
} }
// Convert byte array to base64 const monoData = await ffmpeg.readFile(outputFile);
const rawAudio = btoa(binary); const monoBlob = new Blob([monoData.buffer], { type: 'audio/wav' });
const audioConfig = audioStream.getAudioTracks()[0].getSettings();
const recordData = { const reader = new FileReader();
audio: rawAudio, reader.readAsArrayBuffer(monoBlob);
duration: recordDuration, reader.onload = async (event) => {
channels: audioConfig.channelCount, const arrayBuffer = event.target.result;
sampleRate: audioConfig.sampleRate, const audioContext = new AudioContext();
sampleSize: audioConfig.sampleSize, const arrayBufferCopy = arrayBuffer.slice(0);
}; const audioBufferDecoded = await audioContext.decodeAudioData(arrayBufferCopy);
const audioData = audioBufferDecoded.getChannelData(0);
const audioArray = Array.from(audioData);
const result = genFingerprint(audioArray, audioBufferDecoded.sampleRate);
if (result.error !== 0) {
toast["error"](() => <div>An error occured</div>)
console.log("An error occured: ", result)
return
}
const fingerprintMap = result.data.reduce((dict, item) => {
dict[item.address] = item.anchorTime;
return dict;
}, {});
if (sendRecordingRef.current) { if (sendRecordingRef.current) {
socket.emit("newFingerprint", JSON.stringify({ fingerprint: fingerprintMap }));
}
if (uploadRecording) {
var bytes = new Uint8Array(arrayBuffer);
var rawAudio = "";
for (var i = 0; i < bytes.byteLength; i++) {
rawAudio += String.fromCharCode(bytes[i]);
}
const dataView = new DataView(arrayBuffer);
const recordData = {
audio: btoa(rawAudio),
channels: dataView.getUint16(22, true),
sampleRate: dataView.getUint16(24, true),
sampleSize: dataView.getUint16(34, true),
duration: audioBufferDecoded.duration,
};
console.log("Record data: ", recordData);
socket.emit("newRecording", JSON.stringify(recordData)); socket.emit("newRecording", JSON.stringify(recordData));
} }
}; };
@ -207,10 +250,11 @@ function App() {
} }
} }
function downloadRecording(blob) { function downloadRecording(blob) {
const blobUrl = URL.createObjectURL(blob); const blobUrl = URL.createObjectURL(blob);
// Create a download link
const downloadLink = document.createElement("a"); const downloadLink = document.createElement("a");
downloadLink.href = blobUrl; downloadLink.href = blobUrl;
downloadLink.download = "recorded_audio.wav"; downloadLink.download = "recorded_audio.wav";
@ -244,7 +288,7 @@ function App() {
return ( return (
<div className="App"> <div className="App">
<div className="TopHeader"> <div className="TopHeader">
<h2 style={{ color: "#374151" }}>SeekTune</h2> <h2 style={{ color: "#374151" }}>!Shazam</h2>
<h4 style={{ display: "flex", justifyContent: "flex-end" }}> <h4 style={{ display: "flex", justifyContent: "flex-end" }}>
<AnimatedNumber includeComma={true} animateToNumber={totalSongs} /> <AnimatedNumber includeComma={true} animateToNumber={totalSongs} />
&nbsp;Songs &nbsp;Songs