Tired of digital clutter? Overwhelmed by disorganized files scattered across your computer? Let AI do the heavy lifting! The Local File Organizer is your personal organizing assistant, using cutting-edge AI to bring order to your file chaos - all while respecting your privacy.
Before:
/home/user/messy_documents/
βββ IMG_20230515_140322.jpg
βββ IMG_20230516_083045.jpg
βββ IMG_20230517_192130.jpg
βββ budget_2023.xlsx
βββ meeting_notes_05152023.txt
βββ project_proposal_draft.docx
βββ random_thoughts.txt
βββ recipe_chocolate_cake.pdf
βββ scan0001.pdf
βββ vacation_itinerary.docx
βββ work_presentation.pptx
0 directories, 11 files
After:
/home/user/organized_documents/
βββ Financial
β βββ 2023_Budget_Spreadsheet.xlsx
βββ Food_and_Recipes
β βββ Chocolate_Cake_Recipe.pdf
βββ Meetings_and_Notes
β βββ Team_Meeting_Notes_May_15_2023.txt
βββ Personal
β βββ Random_Thoughts_and_Ideas.txt
βββ Photos
β βββ Cityscape_Sunset_May_17_2023.jpg
β βββ Morning_Coffee_Shop_May_16_2023.jpg
β βββ Office_Team_Lunch_May_15_2023.jpg
βββ Travel
β βββ Summer_Vacation_Itinerary_2023.docx
βββ Work
βββ Project_X_Proposal_Draft.docx
βββ Quarterly_Sales_Report.pdf
βββ Marketing_Strategy_Presentation.pptx
7 directories, 11 files
[2024/09] v0.0.2:
- Featured by Nexa Gallery and Nexa SDK Cookbook!
- Dry Run Mode: check sorting results before committing changes
- Silent Mode: save all logs to a txt file for quieter operation
- Added file support:
.md
, .excel
,.ppt
, and.csv
- Three sorting options: by content, by date, and by type
- The default text model is now Llama3.2 3B
- Improved CLI interaction experience
- Added real-time progress bar for file analysis
Please update the project by deleting the original project folder and reinstalling the requirements. Refer to the installation guide from Step 4.
- Copilot Mode: chat with AI to tell AI how you want to sort the file (ie. read and rename all the PDFs)
- Change models with CLI
- ebook format support
- audio file support
- video file support
- Implement best practices like Johnny Decimal
- Check file duplication
- Dockerfile for easier installation
- People from Nexa is helping me to make executables for macOS, Linux and Windows
This intelligent file organizer harnesses the power of advanced AI models, including language models (LMs) and vision-language models (VLMs), to automate the process of organizing files by:
-
Scanning a specified input directory for files.
-
Content Understanding:
- Textual Analysis: Uses the Llama3.2 3B to analyze and summarize text-based content, generating relevant descriptions and filenames.
- Visual Content Analysis: Uses the LLaVA-v1.6 , based on Vicuna-7B, to interpret visual files such as images, providing context-aware categorization and descriptions.
-
Understanding the content of your files (text, images, and more) to generate relevant descriptions, folder names, and filenames.
-
Organizing the files into a new directory structure based on the generated metadata.
The best part? All AI processing happens 100% on your local device using the Nexa SDK. No internet connection required, no data leaves your computer, and no AI API is needed - keeping your files completely private and secure.
- Images:
.png
,.jpg
,.jpeg
,.gif
,.bmp
- Text Files:
.txt
,.docx
,.md
- Spreadsheets:
.xlsx
,.csv
- Presentations:
.ppt
,.pptx
- PDFs:
.pdf
- Operating System: Compatible with Windows, macOS, and Linux.
- Python Version: Python 3.12
- Conda: Anaconda or Miniconda installed.
- Git: For cloning the repository (or you can download the code as a ZIP file).
For SDK installation and model-related issues, please post on here.
Before installing the Local File Organizer, make sure you have Python installed on your system. We recommend using Python 3.12 or later.
You can download Python from the official website.
Follow the installation instructions for your operating system.
Clone this repository to your local machine using Git:
git clone https://github.com/QiuYannnn/Local-File-Organizer.git
Or download the repository as a ZIP file and extract it to your desired location.
Create a new Conda environment named local_file_organizer
with Python 3.12:
conda create --name local_file_organizer python=3.12
Activate the environment:
conda activate local_file_organizer
To install the CPU version of Nexa SDK, run:
pip install nexaai --prefer-binary --index-url https://nexaai.github.io/nexa-sdk/whl/cpu --extra-index-url https://pypi.org/simple --no-cache-dir
For the GPU version supporting Metal (macOS), run:
CMAKE_ARGS="-DGGML_METAL=ON -DSD_METAL=ON" pip install nexaai --prefer-binary --index-url https://nexaai.github.io/nexa-sdk/whl/metal --extra-index-url https://pypi.org/simple --no-cache-dir
For detailed installation instructions of Nexa SDK for CUDA and AMD GPU support, please refer to the Installation section in the main README.
-
Ensure you are in the project directory:
cd path/to/Local-File-Organizer
Replace
path/to/Local-File-Organizer
with the actual path where you cloned or extracted the project. -
Install the required dependencies:
pip install -r requirements.txt
Note: If you encounter issues with any packages, install them individually:
pip install nexa Pillow pytesseract PyMuPDF python-docx
With the environment activated and dependencies installed, run the script using:
python main.py
-
SDK Models:
- The script uses
NexaVLMInference
andNexaTextInference
models usage. - Ensure you have access to these models and they are correctly set up.
- You may need to download model files or configure paths.
- The script uses
-
Dependencies:
- pytesseract: Requires Tesseract OCR installed on your system.
- macOS:
brew install tesseract
- Ubuntu/Linux:
sudo apt-get install tesseract-ocr
- Windows: Download from Tesseract OCR Windows Installer
- macOS:
- PyMuPDF (fitz): Used for reading PDFs.
- pytesseract: Requires Tesseract OCR installed on your system.
-
Processing Time:
- Processing may take time depending on the number and size of files.
- The script uses multiprocessing to improve performance.
-
Customizing Prompts:
- You can adjust prompts in
data_processing.py
to change how metadata is generated.
- You can adjust prompts in
This project is dual-licensed under the MIT License and Apache 2.0 License. You may choose which license you prefer to use for this project.
- See the MIT License for more details.