Aim of this project is to work seamlessly to decode the speech of two speakers simultaneously in real-time.
LiveLipNet-Duo presents a novel approach in assistive technologies for overcoming the com- munication barriers for the individuals with hearing impairments. Our model leverages the synergy between the spatiotemporal convolutions and recurrent neural network architectures to transcribe the real-time interpretation of speech from two speakers simultaneously. For the elimnation of the manual segmentation, we employ the connectionist temporal classification (CTC) loss which enables direct, end-to-end training from unprocessed video input to textual tran- scription. This project aims not only to improve accessibility for those with hearing impairments, but also extends the use of lipreading into loud environments, where other speech recognition systems that rely solely on audio are ineffective.
This guide will walk you through setting up the Python environment required to run the projects in the Lip_reading repository.
Ensure you have Python installed on your system. You can download Python from python.org.
First, clone the LiveLipNet-Duo repository to your local machine:
git clone https://github.com/33ron33/Lip_reading.git
Create a virtual environment nmaes 'LiveLipNet-env' within the repository directory:
python -m venv LiveLipNet-env
Activate the Virtual Environment
- On macOS/Linux:
source LiveLipNet-env/bin/activate
- On Windows:
LiveLipNet-env\Scripts\activate
Install the project dependencies from the 'requirements.txt' file:
pip install -r requirements.txt
Verify Dependencies:
pip list
To set up the IPython kernel for the virtual environment, run the following command:
python -m ipykernel install --name=LiveLipNet-env
deactivate